Dec 07 09:06:22 crc systemd[1]: Starting Kubernetes Kubelet... Dec 07 09:06:22 crc restorecon[4635]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:22 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 09:06:23 crc restorecon[4635]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 07 09:06:23 crc kubenswrapper[4838]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 09:06:23 crc kubenswrapper[4838]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 07 09:06:23 crc kubenswrapper[4838]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 09:06:23 crc kubenswrapper[4838]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 09:06:23 crc kubenswrapper[4838]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 07 09:06:23 crc kubenswrapper[4838]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.452101 4838 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457748 4838 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457781 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457789 4838 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457796 4838 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457802 4838 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457809 4838 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457851 4838 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457857 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457864 4838 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457873 4838 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457889 4838 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457896 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457902 4838 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457908 4838 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457914 4838 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457921 4838 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457927 4838 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457933 4838 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457939 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457945 4838 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457951 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457958 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457969 4838 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457975 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457981 4838 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457987 4838 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457993 4838 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.457999 4838 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458005 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458011 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458018 4838 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458024 4838 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458034 4838 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458044 4838 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458053 4838 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458065 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458079 4838 feature_gate.go:330] unrecognized feature gate: Example Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458086 4838 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458092 4838 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458098 4838 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458105 4838 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458111 4838 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458117 4838 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458123 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458129 4838 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458138 4838 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458145 4838 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458158 4838 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458165 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458171 4838 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458177 4838 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458183 4838 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458192 4838 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458198 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458203 4838 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458209 4838 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458216 4838 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458222 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458228 4838 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458239 4838 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458245 4838 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458251 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458257 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458264 4838 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458269 4838 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458275 4838 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458281 4838 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458287 4838 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458297 4838 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458305 4838 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.458317 4838 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459025 4838 flags.go:64] FLAG: --address="0.0.0.0" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459065 4838 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459096 4838 flags.go:64] FLAG: --anonymous-auth="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459108 4838 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459139 4838 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459149 4838 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459169 4838 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459180 4838 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459189 4838 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459198 4838 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459208 4838 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459218 4838 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459227 4838 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459236 4838 flags.go:64] FLAG: --cgroup-root="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459244 4838 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459253 4838 flags.go:64] FLAG: --client-ca-file="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459262 4838 flags.go:64] FLAG: --cloud-config="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459271 4838 flags.go:64] FLAG: --cloud-provider="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459280 4838 flags.go:64] FLAG: --cluster-dns="[]" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459292 4838 flags.go:64] FLAG: --cluster-domain="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459300 4838 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459310 4838 flags.go:64] FLAG: --config-dir="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459319 4838 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459329 4838 flags.go:64] FLAG: --container-log-max-files="5" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459340 4838 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459350 4838 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459360 4838 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459369 4838 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459378 4838 flags.go:64] FLAG: --contention-profiling="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459387 4838 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459396 4838 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459406 4838 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459415 4838 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459427 4838 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459437 4838 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459447 4838 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459456 4838 flags.go:64] FLAG: --enable-load-reader="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459465 4838 flags.go:64] FLAG: --enable-server="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459474 4838 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459485 4838 flags.go:64] FLAG: --event-burst="100" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459495 4838 flags.go:64] FLAG: --event-qps="50" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459504 4838 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459523 4838 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459532 4838 flags.go:64] FLAG: --eviction-hard="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459542 4838 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459551 4838 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459560 4838 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459569 4838 flags.go:64] FLAG: --eviction-soft="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459578 4838 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459587 4838 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459596 4838 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459605 4838 flags.go:64] FLAG: --experimental-mounter-path="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459614 4838 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459623 4838 flags.go:64] FLAG: --fail-swap-on="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459632 4838 flags.go:64] FLAG: --feature-gates="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459642 4838 flags.go:64] FLAG: --file-check-frequency="20s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459651 4838 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459661 4838 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459670 4838 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459679 4838 flags.go:64] FLAG: --healthz-port="10248" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459688 4838 flags.go:64] FLAG: --help="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459697 4838 flags.go:64] FLAG: --hostname-override="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459705 4838 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459716 4838 flags.go:64] FLAG: --http-check-frequency="20s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459726 4838 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459738 4838 flags.go:64] FLAG: --image-credential-provider-config="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459747 4838 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459756 4838 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459765 4838 flags.go:64] FLAG: --image-service-endpoint="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459773 4838 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459782 4838 flags.go:64] FLAG: --kube-api-burst="100" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459791 4838 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459800 4838 flags.go:64] FLAG: --kube-api-qps="50" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459810 4838 flags.go:64] FLAG: --kube-reserved="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459842 4838 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459851 4838 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459860 4838 flags.go:64] FLAG: --kubelet-cgroups="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459869 4838 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459878 4838 flags.go:64] FLAG: --lock-file="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459887 4838 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459896 4838 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459905 4838 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459918 4838 flags.go:64] FLAG: --log-json-split-stream="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459929 4838 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459939 4838 flags.go:64] FLAG: --log-text-split-stream="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459949 4838 flags.go:64] FLAG: --logging-format="text" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459958 4838 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459967 4838 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459976 4838 flags.go:64] FLAG: --manifest-url="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459985 4838 flags.go:64] FLAG: --manifest-url-header="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.459997 4838 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460006 4838 flags.go:64] FLAG: --max-open-files="1000000" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460017 4838 flags.go:64] FLAG: --max-pods="110" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460025 4838 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460035 4838 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460044 4838 flags.go:64] FLAG: --memory-manager-policy="None" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460052 4838 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460062 4838 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460071 4838 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460081 4838 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460103 4838 flags.go:64] FLAG: --node-status-max-images="50" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460112 4838 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460122 4838 flags.go:64] FLAG: --oom-score-adj="-999" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460131 4838 flags.go:64] FLAG: --pod-cidr="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460140 4838 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460152 4838 flags.go:64] FLAG: --pod-manifest-path="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460161 4838 flags.go:64] FLAG: --pod-max-pids="-1" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460170 4838 flags.go:64] FLAG: --pods-per-core="0" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460178 4838 flags.go:64] FLAG: --port="10250" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460188 4838 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460196 4838 flags.go:64] FLAG: --provider-id="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460205 4838 flags.go:64] FLAG: --qos-reserved="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460214 4838 flags.go:64] FLAG: --read-only-port="10255" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460223 4838 flags.go:64] FLAG: --register-node="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460231 4838 flags.go:64] FLAG: --register-schedulable="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460240 4838 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460254 4838 flags.go:64] FLAG: --registry-burst="10" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460280 4838 flags.go:64] FLAG: --registry-qps="5" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460289 4838 flags.go:64] FLAG: --reserved-cpus="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460301 4838 flags.go:64] FLAG: --reserved-memory="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460312 4838 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460321 4838 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460330 4838 flags.go:64] FLAG: --rotate-certificates="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460339 4838 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460347 4838 flags.go:64] FLAG: --runonce="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460356 4838 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460365 4838 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460374 4838 flags.go:64] FLAG: --seccomp-default="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460383 4838 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460392 4838 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460402 4838 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460412 4838 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460421 4838 flags.go:64] FLAG: --storage-driver-password="root" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460429 4838 flags.go:64] FLAG: --storage-driver-secure="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460438 4838 flags.go:64] FLAG: --storage-driver-table="stats" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460447 4838 flags.go:64] FLAG: --storage-driver-user="root" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460456 4838 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460466 4838 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460475 4838 flags.go:64] FLAG: --system-cgroups="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460484 4838 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460498 4838 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460507 4838 flags.go:64] FLAG: --tls-cert-file="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460516 4838 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460527 4838 flags.go:64] FLAG: --tls-min-version="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460536 4838 flags.go:64] FLAG: --tls-private-key-file="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460544 4838 flags.go:64] FLAG: --topology-manager-policy="none" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460554 4838 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460562 4838 flags.go:64] FLAG: --topology-manager-scope="container" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460571 4838 flags.go:64] FLAG: --v="2" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460582 4838 flags.go:64] FLAG: --version="false" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460593 4838 flags.go:64] FLAG: --vmodule="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460604 4838 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.460613 4838 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460847 4838 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460859 4838 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460871 4838 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460882 4838 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460893 4838 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460903 4838 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460912 4838 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460920 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460929 4838 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460937 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460945 4838 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460953 4838 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460961 4838 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460970 4838 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460978 4838 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460986 4838 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.460994 4838 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461002 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461011 4838 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461020 4838 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461028 4838 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461036 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461043 4838 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461051 4838 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461059 4838 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461067 4838 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461075 4838 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461083 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461090 4838 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461098 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461106 4838 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461114 4838 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461122 4838 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461129 4838 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461137 4838 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461145 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461152 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461160 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461168 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461177 4838 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461186 4838 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461196 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461204 4838 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461211 4838 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461219 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461227 4838 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461235 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461243 4838 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461251 4838 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461258 4838 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461266 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461274 4838 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461281 4838 feature_gate.go:330] unrecognized feature gate: Example Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461289 4838 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461297 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461305 4838 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461313 4838 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461320 4838 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461328 4838 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461338 4838 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461349 4838 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461357 4838 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461365 4838 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461376 4838 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461385 4838 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461394 4838 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461403 4838 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461411 4838 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461420 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461428 4838 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.461435 4838 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.461460 4838 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.469754 4838 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.469799 4838 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.469971 4838 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.469997 4838 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470007 4838 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470017 4838 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470027 4838 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470038 4838 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470047 4838 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470057 4838 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470068 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470079 4838 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470089 4838 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470099 4838 feature_gate.go:330] unrecognized feature gate: Example Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470109 4838 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470119 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470128 4838 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470138 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470148 4838 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470157 4838 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470167 4838 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470176 4838 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470185 4838 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470193 4838 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470200 4838 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470208 4838 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470215 4838 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470223 4838 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470232 4838 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470239 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470247 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470255 4838 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470263 4838 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470271 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470279 4838 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470287 4838 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470301 4838 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470313 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470323 4838 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470332 4838 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470344 4838 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470353 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470362 4838 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470370 4838 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470378 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470386 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470396 4838 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470404 4838 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470411 4838 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470419 4838 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470427 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470435 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470443 4838 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470451 4838 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470459 4838 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470467 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470475 4838 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470484 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470492 4838 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470500 4838 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470507 4838 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470515 4838 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470523 4838 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470531 4838 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470539 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470547 4838 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470555 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470563 4838 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470571 4838 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470578 4838 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470587 4838 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470597 4838 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470609 4838 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.470632 4838 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470891 4838 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470905 4838 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470915 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470924 4838 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470932 4838 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470941 4838 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470950 4838 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470958 4838 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470967 4838 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470975 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470983 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.470992 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471000 4838 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471008 4838 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471016 4838 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471023 4838 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471032 4838 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471042 4838 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471052 4838 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471061 4838 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471070 4838 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471079 4838 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471087 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471095 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471103 4838 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471113 4838 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471122 4838 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471132 4838 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471142 4838 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471152 4838 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471162 4838 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471172 4838 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471182 4838 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471192 4838 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471204 4838 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471215 4838 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471224 4838 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471235 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471247 4838 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471258 4838 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471269 4838 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471279 4838 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471289 4838 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471300 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471309 4838 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471319 4838 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471331 4838 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471342 4838 feature_gate.go:330] unrecognized feature gate: Example Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471353 4838 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471362 4838 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471370 4838 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471377 4838 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471385 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471393 4838 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471401 4838 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471408 4838 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471416 4838 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471424 4838 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471432 4838 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471442 4838 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471451 4838 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471459 4838 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471466 4838 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471475 4838 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471482 4838 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471489 4838 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471497 4838 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471508 4838 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471519 4838 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471530 4838 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.471542 4838 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.471555 4838 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.472073 4838 server.go:940] "Client rotation is on, will bootstrap in background" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.476985 4838 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.477141 4838 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.478006 4838 server.go:997] "Starting client certificate rotation" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.478044 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.478516 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-18 08:04:57.325675415 +0000 UTC Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.478609 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.485205 4838 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.487381 4838 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.488685 4838 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.496732 4838 log.go:25] "Validated CRI v1 runtime API" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.511882 4838 log.go:25] "Validated CRI v1 image API" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.513295 4838 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.515568 4838 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-07-09-00-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.515594 4838 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.528798 4838 manager.go:217] Machine: {Timestamp:2025-12-07 09:06:23.527727313 +0000 UTC m=+0.235046350 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:db13a37f-60ef-4ad4-906a-e0d1f4e45870 BootID:98e1eaea-301b-4b2e-acd1-6718eb434b9d Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b7:fe:1e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b7:fe:1e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d7:77:eb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a0:ee:6b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:30:cc:58 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:43:23:d9 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:46:b8:fa Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:46:28:1a:b7:a2 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:aa:55:f6:fb:24:f8 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.529002 4838 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.529176 4838 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.529752 4838 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.530021 4838 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.530072 4838 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.530320 4838 topology_manager.go:138] "Creating topology manager with none policy" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.530333 4838 container_manager_linux.go:303] "Creating device plugin manager" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.530590 4838 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.530640 4838 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.530973 4838 state_mem.go:36] "Initialized new in-memory state store" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.531083 4838 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.532443 4838 kubelet.go:418] "Attempting to sync node with API server" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.532467 4838 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.532496 4838 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.532512 4838 kubelet.go:324] "Adding apiserver pod source" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.532526 4838 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.534466 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.534615 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.534643 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.534724 4838 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.534841 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.535172 4838 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.535864 4838 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536405 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536432 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536441 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536449 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536462 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536470 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536478 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536493 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536504 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536513 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536524 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536533 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.536741 4838 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.537210 4838 server.go:1280] "Started kubelet" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.537686 4838 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.537938 4838 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.537665 4838 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.538363 4838 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.539112 4838 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.539144 4838 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.539729 4838 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 13:22:59.096264673 +0000 UTC Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.539762 4838 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 196h16m35.556504973s for next certificate rotation Dec 07 09:06:23 crc systemd[1]: Started Kubernetes Kubelet. Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.540301 4838 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.540315 4838 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.540431 4838 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.541297 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.541750 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.541643 4838 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.73:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ee3a0c059a312 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-07 09:06:23.537185554 +0000 UTC m=+0.244504571,LastTimestamp:2025-12-07 09:06:23.537185554 +0000 UTC m=+0.244504571,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.542118 4838 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.542164 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="200ms" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.544371 4838 factory.go:55] Registering systemd factory Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.544460 4838 factory.go:221] Registration of the systemd container factory successfully Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.544770 4838 factory.go:153] Registering CRI-O factory Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.544879 4838 factory.go:221] Registration of the crio container factory successfully Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.544996 4838 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.545099 4838 factory.go:103] Registering Raw factory Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.545177 4838 manager.go:1196] Started watching for new ooms in manager Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.544927 4838 server.go:460] "Adding debug handlers to kubelet server" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.546003 4838 manager.go:319] Starting recovery of all containers Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564456 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564543 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564576 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564606 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564635 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564662 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564690 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564717 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564749 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564774 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564802 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564881 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564914 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564949 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.564979 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565028 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565092 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565155 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565194 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565227 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565255 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565285 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565313 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565344 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565371 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565397 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565433 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565461 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565490 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565529 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565554 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565582 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565607 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565633 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565659 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565689 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565719 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565747 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565775 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565801 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565913 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565946 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.565975 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566037 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566068 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566095 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566123 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566175 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566208 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566240 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566269 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566341 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566382 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566412 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566440 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566471 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566499 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566530 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566555 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566584 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566615 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566642 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566670 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566696 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566725 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566753 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566780 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566806 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566907 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566966 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.566994 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567021 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567050 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567077 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567104 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567133 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567165 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567193 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567222 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567248 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567276 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567302 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567330 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567357 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567386 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567414 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567444 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567472 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567696 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567723 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567752 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567780 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567844 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567878 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567906 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567934 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567965 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.567995 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568022 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568049 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568079 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568107 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568163 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568192 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568235 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568266 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568298 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568329 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568364 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568395 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568425 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568455 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568484 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568708 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568736 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568763 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568792 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568854 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568884 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568913 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568945 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.568971 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569000 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569024 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569053 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569082 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569108 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569133 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569159 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569186 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569211 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569238 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.569266 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570573 4838 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570635 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570665 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570694 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570722 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570751 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570777 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570806 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570880 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570910 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570940 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570966 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.570994 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571022 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571053 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571082 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571109 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571137 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571217 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571247 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571275 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571306 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571334 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571364 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571392 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571419 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571446 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571475 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571505 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571533 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571561 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571603 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571629 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571661 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571686 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571717 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571744 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571769 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571794 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571859 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571889 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571914 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571939 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571965 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.571992 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572017 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572041 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572068 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572092 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572145 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572176 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572206 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572233 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572258 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572284 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572310 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572332 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572351 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572367 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572385 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572402 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572419 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572438 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572456 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572474 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572493 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572512 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572530 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572549 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572568 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572588 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572608 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572627 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572646 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572666 4838 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572681 4838 reconstruct.go:97] "Volume reconstruction finished" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.572692 4838 reconciler.go:26] "Reconciler: start to sync state" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.590955 4838 manager.go:324] Recovery completed Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.600762 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.603189 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.603228 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.603253 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.606538 4838 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.606550 4838 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.606635 4838 state_mem.go:36] "Initialized new in-memory state store" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.611497 4838 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.613244 4838 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.613282 4838 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.613307 4838 kubelet.go:2335] "Starting kubelet main sync loop" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.613341 4838 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.643625 4838 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.713479 4838 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 07 09:06:23 crc kubenswrapper[4838]: W1207 09:06:23.730072 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.730153 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.744114 4838 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.744708 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="400ms" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.829248 4838 policy_none.go:49] "None policy: Start" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.831087 4838 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.831170 4838 state_mem.go:35] "Initializing new in-memory state store" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.845466 4838 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.913888 4838 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.946139 4838 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.954415 4838 manager.go:334] "Starting Device Plugin manager" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.954509 4838 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.954530 4838 server.go:79] "Starting device plugin registration server" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.955122 4838 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.955154 4838 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.955443 4838 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.955568 4838 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 07 09:06:23 crc kubenswrapper[4838]: I1207 09:06:23.955597 4838 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 07 09:06:23 crc kubenswrapper[4838]: E1207 09:06:23.966589 4838 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.055715 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.057262 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.057482 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.057616 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.057851 4838 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.058779 4838 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.146761 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="800ms" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.259803 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.261437 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.261639 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.261767 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.261955 4838 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.262663 4838 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.314890 4838 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.315470 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.317630 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.317701 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.317720 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.317957 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.318484 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.318573 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.319324 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.319377 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.319447 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.320123 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.320118 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.320482 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.320508 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.321025 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.321275 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.322314 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.322371 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.322393 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.322570 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.323122 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.323443 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.324002 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.324054 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.324073 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.324691 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.324971 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.325167 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.325551 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.326707 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.326984 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.327855 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.327905 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.327921 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.327954 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.327928 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.327973 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.328982 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.329178 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.331037 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.331080 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.331097 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.331252 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.331272 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.331286 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383047 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383110 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383173 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383211 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383244 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383276 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383306 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383335 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383365 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383416 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383462 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383501 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383546 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383587 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.383627 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485009 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485090 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485124 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485156 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485187 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485204 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485218 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485285 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485325 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485356 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485353 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485403 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485378 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485378 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485472 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485534 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485550 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485581 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485585 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485636 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485673 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485739 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485751 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485811 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485805 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485888 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485891 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485900 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.485951 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.486004 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.539897 4838 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.658945 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.663703 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.664684 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.664719 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.664728 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.664750 4838 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.665358 4838 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Dec 07 09:06:24 crc kubenswrapper[4838]: W1207 09:06:24.678852 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.678920 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:24 crc kubenswrapper[4838]: W1207 09:06:24.679157 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.679236 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:24 crc kubenswrapper[4838]: W1207 09:06:24.683020 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-55e9f30641cc4786f87fade6fa496cd55c47ded1e9e69ca4937b0c59441ba3fd WatchSource:0}: Error finding container 55e9f30641cc4786f87fade6fa496cd55c47ded1e9e69ca4937b0c59441ba3fd: Status 404 returned error can't find the container with id 55e9f30641cc4786f87fade6fa496cd55c47ded1e9e69ca4937b0c59441ba3fd Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.684338 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.694252 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.725718 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: I1207 09:06:24.737477 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:24 crc kubenswrapper[4838]: W1207 09:06:24.753429 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-2569d6af7ad1537817e8ebc9ccda4041911eb16741506c0c00b5241e97d0917d WatchSource:0}: Error finding container 2569d6af7ad1537817e8ebc9ccda4041911eb16741506c0c00b5241e97d0917d: Status 404 returned error can't find the container with id 2569d6af7ad1537817e8ebc9ccda4041911eb16741506c0c00b5241e97d0917d Dec 07 09:06:24 crc kubenswrapper[4838]: W1207 09:06:24.764761 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-524db78180be1be7e4796d9b9c343c1c096069f1815497d603b62e26d9d7713f WatchSource:0}: Error finding container 524db78180be1be7e4796d9b9c343c1c096069f1815497d603b62e26d9d7713f: Status 404 returned error can't find the container with id 524db78180be1be7e4796d9b9c343c1c096069f1815497d603b62e26d9d7713f Dec 07 09:06:24 crc kubenswrapper[4838]: W1207 09:06:24.930606 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.930688 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:24 crc kubenswrapper[4838]: E1207 09:06:24.947514 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="1.6s" Dec 07 09:06:25 crc kubenswrapper[4838]: W1207 09:06:25.072436 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:25 crc kubenswrapper[4838]: E1207 09:06:25.072956 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.465961 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.467516 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.467544 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.467556 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.467582 4838 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 09:06:25 crc kubenswrapper[4838]: E1207 09:06:25.468310 4838 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.73:6443: connect: connection refused" node="crc" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.519433 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 07 09:06:25 crc kubenswrapper[4838]: E1207 09:06:25.520381 4838 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.73:6443: connect: connection refused" logger="UnhandledError" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.539271 4838 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.73:6443: connect: connection refused Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.625917 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42" exitCode=0 Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.625968 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.626062 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"55e9f30641cc4786f87fade6fa496cd55c47ded1e9e69ca4937b0c59441ba3fd"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.626134 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.627201 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.627227 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.627237 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.630872 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5116f0a6912aff4cc613ea60bbb3c2920019e922282834e91033770defcae9c6"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.630908 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9fbb4f752e3f764b7c5c59d91a7965df1fbea45e17d532e43de493d8c4223233"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.630923 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"524db78180be1be7e4796d9b9c343c1c096069f1815497d603b62e26d9d7713f"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.633045 4838 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01" exitCode=0 Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.633097 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.633118 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2569d6af7ad1537817e8ebc9ccda4041911eb16741506c0c00b5241e97d0917d"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.633202 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.635029 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.635067 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.635083 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.639954 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.639902 4838 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27" exitCode=0 Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.640165 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"66fa11234d66ba89c95a5af4b4c606f449d6903535a8c5b54908bb385dfc0300"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.640190 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.641339 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.641371 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.641384 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.650658 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.651431 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.651449 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.651457 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.653928 4838 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3e4842f8940896c6905ac0228724049b0654a57dad7febe56c5cc823ede79ccb" exitCode=0 Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.653997 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3e4842f8940896c6905ac0228724049b0654a57dad7febe56c5cc823ede79ccb"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.654034 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e29e6ca87bce569e0465db20bdf6ec4bdf938cb621797eee5475774f14278a5e"} Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.654174 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.654998 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.655033 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:25 crc kubenswrapper[4838]: I1207 09:06:25.655046 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.668191 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.668245 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.668261 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.668396 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.669273 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.669301 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.669312 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.675651 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.675687 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.675702 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.675716 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.683393 4838 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="be9c7cdfe6ad30105c4cc4f1ed1f840a93ebb6267da0ff4665e6cde3af1ea4de" exitCode=0 Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.683442 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"be9c7cdfe6ad30105c4cc4f1ed1f840a93ebb6267da0ff4665e6cde3af1ea4de"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.683540 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.687243 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.687277 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.687287 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.690902 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"286fc6c21ddb56cdcb4bdb001dbbbcbd294b23971933e39148c05efed583e5b7"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.690987 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.691724 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.691743 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.691751 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.697564 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"730edf2868680ba9cc75bd212a4e9ac94b28e13735b050683c414de1e2c13ed9"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.697597 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"176f6fb5d35ad470e50545dc0e469d41525e1d167213f0d6461b8d34d83fa88d"} Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.697655 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.699118 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.699147 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:26 crc kubenswrapper[4838]: I1207 09:06:26.699155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.068714 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.070018 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.070067 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.070077 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.070107 4838 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.705482 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341"} Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.705742 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.707433 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.707488 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.707527 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.711531 4838 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5840c0821e38ba00a93aec0fcb515241e7bd2270a3803912813f637ea64911bc" exitCode=0 Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.711694 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.711956 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5840c0821e38ba00a93aec0fcb515241e7bd2270a3803912813f637ea64911bc"} Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.712184 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.713848 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.713885 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.713912 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.713925 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.713889 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.714033 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:27 crc kubenswrapper[4838]: I1207 09:06:27.906390 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.502803 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.567340 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.720578 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4c0c0a1569c8f493c94724452ae774d77156757f42a84b26edb8d5b68a91cf24"} Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.720626 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.720651 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"537596a1dbeb1b6b5cc4857605384d3fcb8f3c053a9e3dac23f0a3177fc78c6e"} Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.720673 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"97895fc09c3def5f0c402ca3021ae4101042d0e3525ed7c9d83172d1d188cc50"} Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.720695 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9dab50c0e7254b5bd23d2e952f509936ee09e10fd77a3eee99b76e602588529b"} Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.721494 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.721558 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:28 crc kubenswrapper[4838]: I1207 09:06:28.721583 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.549742 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.722144 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.722374 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.724034 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.724079 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.724102 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.732331 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.735796 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d11761904738801ab9c2ada5d856953a211416b0ffc4af042874705385e6a563"} Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.735874 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.735919 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.737263 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.737319 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.737340 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.738057 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.738107 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:29 crc kubenswrapper[4838]: I1207 09:06:29.738130 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.739533 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.739661 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.739676 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.741976 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.742027 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.742053 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.743109 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.743156 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:30 crc kubenswrapper[4838]: I1207 09:06:30.743176 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:31 crc kubenswrapper[4838]: I1207 09:06:31.742698 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:31 crc kubenswrapper[4838]: I1207 09:06:31.744199 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:31 crc kubenswrapper[4838]: I1207 09:06:31.744240 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:31 crc kubenswrapper[4838]: I1207 09:06:31.744258 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:32 crc kubenswrapper[4838]: I1207 09:06:32.664077 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 07 09:06:32 crc kubenswrapper[4838]: I1207 09:06:32.664326 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:32 crc kubenswrapper[4838]: I1207 09:06:32.665947 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:32 crc kubenswrapper[4838]: I1207 09:06:32.665985 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:32 crc kubenswrapper[4838]: I1207 09:06:32.666005 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:33 crc kubenswrapper[4838]: E1207 09:06:33.966986 4838 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 07 09:06:33 crc kubenswrapper[4838]: I1207 09:06:33.969167 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:33 crc kubenswrapper[4838]: I1207 09:06:33.969427 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:33 crc kubenswrapper[4838]: I1207 09:06:33.971096 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:33 crc kubenswrapper[4838]: I1207 09:06:33.971147 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:33 crc kubenswrapper[4838]: I1207 09:06:33.971167 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:34 crc kubenswrapper[4838]: I1207 09:06:34.654683 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 07 09:06:34 crc kubenswrapper[4838]: I1207 09:06:34.655230 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:34 crc kubenswrapper[4838]: I1207 09:06:34.664713 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:34 crc kubenswrapper[4838]: I1207 09:06:34.664780 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:34 crc kubenswrapper[4838]: I1207 09:06:34.664802 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:35 crc kubenswrapper[4838]: I1207 09:06:35.967112 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:06:35 crc kubenswrapper[4838]: I1207 09:06:35.967373 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:35 crc kubenswrapper[4838]: I1207 09:06:35.969081 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:35 crc kubenswrapper[4838]: I1207 09:06:35.969122 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:35 crc kubenswrapper[4838]: I1207 09:06:35.969139 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:36 crc kubenswrapper[4838]: W1207 09:06:36.524736 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.524847 4838 trace.go:236] Trace[1457587769]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 09:06:26.519) (total time: 10005ms): Dec 07 09:06:36 crc kubenswrapper[4838]: Trace[1457587769]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10004ms (09:06:36.524) Dec 07 09:06:36 crc kubenswrapper[4838]: Trace[1457587769]: [10.005021465s] [10.005021465s] END Dec 07 09:06:36 crc kubenswrapper[4838]: E1207 09:06:36.524873 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.539910 4838 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 07 09:06:36 crc kubenswrapper[4838]: E1207 09:06:36.549478 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.557883 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.558316 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.560591 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.560800 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.561024 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.563810 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.758065 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.759323 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.759935 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:36 crc kubenswrapper[4838]: I1207 09:06:36.759977 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:37 crc kubenswrapper[4838]: W1207 09:06:37.044140 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 07 09:06:37 crc kubenswrapper[4838]: I1207 09:06:37.044242 4838 trace.go:236] Trace[1775013676]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 09:06:27.042) (total time: 10001ms): Dec 07 09:06:37 crc kubenswrapper[4838]: Trace[1775013676]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:06:37.044) Dec 07 09:06:37 crc kubenswrapper[4838]: Trace[1775013676]: [10.001254855s] [10.001254855s] END Dec 07 09:06:37 crc kubenswrapper[4838]: E1207 09:06:37.044276 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 07 09:06:37 crc kubenswrapper[4838]: E1207 09:06:37.070889 4838 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 07 09:06:37 crc kubenswrapper[4838]: W1207 09:06:37.549261 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 07 09:06:37 crc kubenswrapper[4838]: I1207 09:06:37.549387 4838 trace.go:236] Trace[536544560]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 09:06:27.547) (total time: 10001ms): Dec 07 09:06:37 crc kubenswrapper[4838]: Trace[536544560]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:06:37.549) Dec 07 09:06:37 crc kubenswrapper[4838]: Trace[536544560]: [10.001604055s] [10.001604055s] END Dec 07 09:06:37 crc kubenswrapper[4838]: E1207 09:06:37.549419 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 07 09:06:37 crc kubenswrapper[4838]: W1207 09:06:37.766435 4838 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 07 09:06:37 crc kubenswrapper[4838]: I1207 09:06:37.766569 4838 trace.go:236] Trace[1533255376]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 09:06:27.764) (total time: 10001ms): Dec 07 09:06:37 crc kubenswrapper[4838]: Trace[1533255376]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:06:37.766) Dec 07 09:06:37 crc kubenswrapper[4838]: Trace[1533255376]: [10.001947412s] [10.001947412s] END Dec 07 09:06:37 crc kubenswrapper[4838]: E1207 09:06:37.766650 4838 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 07 09:06:38 crc kubenswrapper[4838]: I1207 09:06:38.503315 4838 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 09:06:38 crc kubenswrapper[4838]: I1207 09:06:38.503942 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 07 09:06:39 crc kubenswrapper[4838]: E1207 09:06:39.552032 4838 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 07 09:06:39 crc kubenswrapper[4838]: I1207 09:06:39.558623 4838 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 09:06:39 crc kubenswrapper[4838]: I1207 09:06:39.558696 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 09:06:40 crc kubenswrapper[4838]: I1207 09:06:40.271589 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:40 crc kubenswrapper[4838]: I1207 09:06:40.272768 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:40 crc kubenswrapper[4838]: I1207 09:06:40.272831 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:40 crc kubenswrapper[4838]: I1207 09:06:40.272846 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:40 crc kubenswrapper[4838]: I1207 09:06:40.272875 4838 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 09:06:41 crc kubenswrapper[4838]: I1207 09:06:41.639458 4838 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 07 09:06:41 crc kubenswrapper[4838]: I1207 09:06:41.639517 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.116480 4838 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.513519 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.514603 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.516327 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.516393 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.516417 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.522163 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.773152 4838 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.773219 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.774350 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.774412 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:43 crc kubenswrapper[4838]: I1207 09:06:43.774439 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:43 crc kubenswrapper[4838]: E1207 09:06:43.968080 4838 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.692910 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.693139 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.694722 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.694848 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.694880 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.713454 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.776312 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.777768 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.777844 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:44 crc kubenswrapper[4838]: I1207 09:06:44.777859 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.614067 4838 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 07 09:06:46 crc kubenswrapper[4838]: E1207 09:06:46.618507 4838 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.619523 4838 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.683431 4838 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44140->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.683488 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44140->192.168.126.11:17697: read: connection reset by peer" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.683715 4838 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.683730 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.684137 4838 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44150->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.684291 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44150->192.168.126.11:17697: read: connection reset by peer" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.710992 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.711404 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.713289 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.713330 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.713342 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.717101 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.782336 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.784414 4838 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341" exitCode=255 Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.784499 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341"} Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.784781 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.784780 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.786093 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.786127 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.786140 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.786168 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.786199 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.786211 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.786989 4838 scope.go:117] "RemoveContainer" containerID="27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341" Dec 07 09:06:46 crc kubenswrapper[4838]: I1207 09:06:46.860410 4838 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.547271 4838 apiserver.go:52] "Watching apiserver" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.550687 4838 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.550967 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.551579 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.551648 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.551507 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.551657 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.551768 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.551926 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.552290 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.552373 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.552405 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.553984 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.554161 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.554462 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.555419 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.557425 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.557870 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.557921 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.558360 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.566978 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.570338 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.584325 4838 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.590406 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.604621 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.632447 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.645796 4838 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.661049 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.676781 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.685506 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.692768 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.700672 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727018 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727065 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727082 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727099 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727114 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727130 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727147 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727175 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727191 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727205 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727221 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727235 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727248 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727263 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727517 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727584 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727770 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727886 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727929 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727940 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727958 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.727981 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728003 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728017 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728027 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728018 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728025 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728042 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728053 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728123 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728126 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728169 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728195 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728219 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728242 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728265 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728288 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728310 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728331 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728351 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728372 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728396 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728418 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728439 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728460 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728480 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728500 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728742 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728765 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728786 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728833 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728857 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728881 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728903 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728923 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728944 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728966 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729008 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729029 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729051 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729072 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729094 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729116 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729139 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729160 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729185 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729209 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729234 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729257 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729281 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729303 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729327 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729350 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729374 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729395 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729416 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729437 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729459 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729482 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729504 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729524 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729557 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729581 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729602 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729623 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729646 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729672 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729695 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729717 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729737 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729759 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729780 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729802 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729842 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729865 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729888 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729910 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729934 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729968 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729990 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730013 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730034 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730057 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730078 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730099 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730121 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730149 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730171 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730190 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730209 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730229 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730250 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730270 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730291 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730310 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730330 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730351 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730372 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730394 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730413 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730436 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730460 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730479 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730499 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730521 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730541 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730561 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730651 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730675 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730696 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730717 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730737 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730757 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730776 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730798 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730836 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730859 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730879 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730900 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730920 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730942 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730964 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.730986 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731006 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731027 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731073 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731096 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731118 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731144 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731168 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731188 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731211 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731237 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731263 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731289 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731313 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731334 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731356 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731380 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731402 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731426 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731449 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731471 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731492 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731514 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731535 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731558 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731595 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731619 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731643 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731669 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731690 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731713 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731735 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731757 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731779 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731802 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731844 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731869 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731890 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731915 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731939 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731960 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731983 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732005 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732029 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732051 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732073 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732097 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732121 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732144 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732167 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732190 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732214 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732240 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732262 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732284 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732307 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732330 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732355 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732380 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732407 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732432 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732458 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732499 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732530 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732558 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732586 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732614 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732640 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732667 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732693 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732717 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732738 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732763 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732793 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732860 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732893 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732945 4838 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732961 4838 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732975 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732989 4838 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733004 4838 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733018 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733031 4838 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733044 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733060 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728264 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728414 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728542 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728560 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728640 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728684 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.735497 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728842 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.728995 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729075 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729185 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729242 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729306 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729333 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729428 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729576 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729730 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729944 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.729952 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731201 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731408 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731637 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.731898 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732030 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732058 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732304 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732433 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732523 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732563 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732786 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.732899 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.733140 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733167 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733377 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733653 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.733747 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:06:48.233726873 +0000 UTC m=+24.941045900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.733859 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.734255 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.734303 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.734514 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.734658 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.734699 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.735005 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.735024 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.735139 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.735270 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.735385 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.735944 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.736134 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.736253 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.736335 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.736511 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.736526 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.736787 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.737357 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.738426 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.738775 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.739099 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.739400 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.739439 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.739564 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.739727 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.739846 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.740164 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.740560 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.741861 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.741980 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.742224 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.742651 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.742732 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.742850 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743136 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743357 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743450 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743483 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743586 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743591 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743783 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.743859 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744006 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744062 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744142 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744286 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744386 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744451 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744599 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.744750 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.745164 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.745794 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.746088 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.746128 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.746768 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.747155 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.747331 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.747592 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.747794 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.748047 4838 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.748109 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.750257 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.751506 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.751870 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.753996 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.755316 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.755532 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.755764 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.758985 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.759273 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:48.259239959 +0000 UTC m=+24.966558996 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.759368 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:48.259354012 +0000 UTC m=+24.966673049 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.759764 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.760209 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.760905 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.761125 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.761931 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.762422 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.761807 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.762854 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.762888 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.762909 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.763296 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.763731 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.765154 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.766247 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.766786 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.766969 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.767183 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.767449 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.767844 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.768063 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.768473 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.769235 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.769569 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.769980 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.771692 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.771944 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.771972 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.772052 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.772364 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.772436 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.773539 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.773642 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.773733 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.773757 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.773774 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.773859 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:48.27383583 +0000 UTC m=+24.981154867 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.774332 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.774338 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.774516 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.774710 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.774725 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.774735 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:47 crc kubenswrapper[4838]: E1207 09:06:47.774772 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:48.274762226 +0000 UTC m=+24.982081363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.774928 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775050 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775200 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775209 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775287 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.767090 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775862 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775885 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775916 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.775996 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.776251 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.776543 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.779208 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.779400 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.779649 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.781882 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.782214 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.782322 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.782637 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.784959 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.785142 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.785505 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.787686 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.787757 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.788059 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.788451 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.788587 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.789089 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.789368 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.789524 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.789898 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.790089 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.790549 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.791345 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.791500 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.791733 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.791942 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.791839 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.791914 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792003 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792188 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792275 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792296 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792667 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792756 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792942 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792878 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792905 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792937 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.792997 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.793078 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.793453 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9"} Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.794002 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.794054 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.794421 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.797719 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.806257 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.806782 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.809493 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.814337 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.818674 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.822279 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.825259 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.830699 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.833876 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834005 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834014 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834110 4838 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834130 4838 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834143 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834197 4838 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834212 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834225 4838 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834265 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834281 4838 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834292 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834304 4838 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834317 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834352 4838 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834364 4838 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834374 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834386 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834395 4838 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834405 4838 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834443 4838 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834454 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834465 4838 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834475 4838 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834486 4838 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834521 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834533 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834545 4838 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834556 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834566 4838 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834603 4838 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834615 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834627 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834639 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834649 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834685 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834697 4838 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834709 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834720 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834731 4838 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834765 4838 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834776 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834787 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834798 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834114 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834809 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834876 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834887 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834911 4838 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834924 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834936 4838 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834947 4838 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834957 4838 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834965 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834984 4838 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.834992 4838 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835000 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835008 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835016 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835024 4838 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835032 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835039 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835046 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835054 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835062 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835070 4838 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835078 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835085 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835093 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835102 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835110 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835117 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835125 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835133 4838 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835142 4838 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835150 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835157 4838 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835164 4838 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835171 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835180 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835227 4838 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835236 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835243 4838 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835251 4838 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835259 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835266 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835274 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835282 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835299 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835307 4838 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835314 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835322 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835331 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835339 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835347 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835355 4838 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835363 4838 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835371 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835379 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835386 4838 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835395 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835403 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835410 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835419 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835428 4838 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835435 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835444 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835452 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835460 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835468 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835476 4838 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835484 4838 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835491 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835499 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835508 4838 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835516 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835524 4838 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835531 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835539 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835547 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835555 4838 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835562 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835570 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835577 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835585 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835592 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835600 4838 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835608 4838 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835617 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835625 4838 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835633 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835640 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835648 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835657 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835665 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835674 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835682 4838 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835691 4838 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835699 4838 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835707 4838 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835714 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835722 4838 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835730 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835737 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835745 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835756 4838 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835763 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835771 4838 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835780 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835788 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835795 4838 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835803 4838 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835810 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835835 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835846 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835856 4838 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835866 4838 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835876 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835883 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835891 4838 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835898 4838 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835906 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835913 4838 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835921 4838 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835929 4838 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835938 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835946 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835953 4838 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835961 4838 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835968 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835977 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835984 4838 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835992 4838 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.835999 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836007 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836014 4838 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836021 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836032 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836039 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836047 4838 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836054 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836062 4838 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836071 4838 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836079 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836087 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836095 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836103 4838 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836110 4838 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.836118 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.839635 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.847051 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.883174 4838 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.883784 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.901141 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 09:06:47 crc kubenswrapper[4838]: W1207 09:06:47.910491 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-cfc08e2e8942a5937c0d6ffe36a4060a69143ba54788e328a3b7a4f39d1712a4 WatchSource:0}: Error finding container cfc08e2e8942a5937c0d6ffe36a4060a69143ba54788e328a3b7a4f39d1712a4: Status 404 returned error can't find the container with id cfc08e2e8942a5937c0d6ffe36a4060a69143ba54788e328a3b7a4f39d1712a4 Dec 07 09:06:47 crc kubenswrapper[4838]: I1207 09:06:47.915088 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.238833 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.239302 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:06:49.239287102 +0000 UTC m=+25.946606119 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.339772 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.339839 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.339883 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.339911 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.339914 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.339969 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:49.339955325 +0000 UTC m=+26.047274332 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340007 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340022 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340033 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340065 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:49.340055238 +0000 UTC m=+26.047374255 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340114 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340124 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340133 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340159 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:49.34015017 +0000 UTC m=+26.047469187 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340205 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.340246 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:49.340235583 +0000 UTC m=+26.047554600 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.614151 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:48 crc kubenswrapper[4838]: E1207 09:06:48.614268 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.797939 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e"} Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.797988 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"28208c36ac402de01b44248c57dffbc513e97f2debea85ae999e3ac8d884c1d2"} Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.801883 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181"} Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.801915 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe"} Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.801929 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cfc08e2e8942a5937c0d6ffe36a4060a69143ba54788e328a3b7a4f39d1712a4"} Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.803635 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9a06d1a183a15ffc1185a6d8bd1e6b029915e50b1bb90732081d1188c7dcf5af"} Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.818457 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.835635 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.852802 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.865206 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.879331 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.895910 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.911867 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.931873 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.952275 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:48 crc kubenswrapper[4838]: I1207 09:06:48.983281 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:48Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.006501 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:49Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.025110 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:49Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.036685 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:49Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.047000 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:49Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.251112 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.251312 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:06:51.251286965 +0000 UTC m=+27.958606002 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.352503 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.352546 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.352572 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352669 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.352751 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352805 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:51.352787942 +0000 UTC m=+28.060106959 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352689 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352850 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352865 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352899 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:51.352888535 +0000 UTC m=+28.060207562 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352739 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352920 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352930 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.352954 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:51.352946136 +0000 UTC m=+28.060265153 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.353010 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.353042 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:51.353031249 +0000 UTC m=+28.060350266 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.614505 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.614575 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.614673 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:06:49 crc kubenswrapper[4838]: E1207 09:06:49.614751 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.618082 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.618682 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.619719 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.620591 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.621393 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.623205 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.623987 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.625190 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.625854 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.626772 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.627306 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.628386 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.628962 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.629640 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.630896 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.631568 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.632793 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.633469 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.634344 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.635660 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.636280 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.637565 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.638226 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.639552 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.640149 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.641052 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.642764 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.643405 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.644594 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.645266 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.646772 4838 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.647000 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.649071 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.650622 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.651199 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.653385 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.654454 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.655681 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.656526 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.657934 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.658575 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.659965 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.660762 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.662070 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.662674 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.663930 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.664778 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.666332 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.667000 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.668031 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.668627 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.670051 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.670771 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 07 09:06:49 crc kubenswrapper[4838]: I1207 09:06:49.671404 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.613985 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:50 crc kubenswrapper[4838]: E1207 09:06:50.614990 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.810160 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931"} Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.828334 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:50Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.849303 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:50Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.862372 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:50Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.881374 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:50Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.896365 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:50Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.915498 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:50Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:50 crc kubenswrapper[4838]: I1207 09:06:50.929660 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:50Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:51 crc kubenswrapper[4838]: I1207 09:06:51.267205 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.267505 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:06:55.267461893 +0000 UTC m=+31.974780910 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:06:51 crc kubenswrapper[4838]: I1207 09:06:51.368721 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:51 crc kubenswrapper[4838]: I1207 09:06:51.369001 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:51 crc kubenswrapper[4838]: I1207 09:06:51.369097 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:51 crc kubenswrapper[4838]: I1207 09:06:51.369218 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.368996 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.369340 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.369364 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.369082 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.369525 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.369445 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:55.369420374 +0000 UTC m=+32.076739431 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.369651 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:55.36962382 +0000 UTC m=+32.076942877 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.369678 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:55.369664751 +0000 UTC m=+32.076983808 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.370040 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.370087 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.370106 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.370187 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:55.370168065 +0000 UTC m=+32.077487122 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:51 crc kubenswrapper[4838]: I1207 09:06:51.614111 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.614286 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:06:51 crc kubenswrapper[4838]: I1207 09:06:51.614946 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:51 crc kubenswrapper[4838]: E1207 09:06:51.615102 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:06:52 crc kubenswrapper[4838]: I1207 09:06:52.613593 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:52 crc kubenswrapper[4838]: E1207 09:06:52.613971 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.019346 4838 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.020828 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.020866 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.020876 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.020920 4838 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.028316 4838 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.028646 4838 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.029934 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.029968 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.029979 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.029993 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.030003 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.052776 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.056974 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.057167 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.057295 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.057414 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.057511 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.072407 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.076495 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.076543 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.076558 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.076575 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.076588 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.096766 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.100299 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.100348 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.100364 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.100382 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.100393 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.116062 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.120098 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.120152 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.120169 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.120187 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.120201 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.132761 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.132894 4838 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.134404 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.134507 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.134519 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.134536 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.134547 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.236731 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.236803 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.237060 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.237087 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.237358 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.340708 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.340755 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.340775 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.340798 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.340905 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.443781 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.443868 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.443894 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.443921 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.443940 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.546145 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.546188 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.546200 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.546217 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.546229 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.581801 4838 csr.go:261] certificate signing request csr-lcwz5 is approved, waiting to be issued Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.594670 4838 csr.go:257] certificate signing request csr-lcwz5 is issued Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.613971 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.614033 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.614258 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:06:53 crc kubenswrapper[4838]: E1207 09:06:53.614337 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.632213 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.648107 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.648335 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.648417 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.648484 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.648542 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.666443 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.684715 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.693227 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-b26td"] Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.693650 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-rf7dv"] Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.693829 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.694019 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.697077 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.697425 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.697477 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.697577 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.697493 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.697493 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.697494 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.712150 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.729475 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.741418 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.750759 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.750791 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.750801 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.750839 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.750853 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.757954 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.773080 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.789960 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kt2d\" (UniqueName: \"kubernetes.io/projected/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-kube-api-access-2kt2d\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.790012 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/09bf1dd6-0f35-4ec3-9d94-87800241f5eb-hosts-file\") pod \"node-resolver-rf7dv\" (UID: \"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\") " pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.790034 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xx2m\" (UniqueName: \"kubernetes.io/projected/09bf1dd6-0f35-4ec3-9d94-87800241f5eb-kube-api-access-5xx2m\") pod \"node-resolver-rf7dv\" (UID: \"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\") " pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.790064 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-serviceca\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.790093 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-host\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.796091 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.812197 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.824182 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.833713 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.847318 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.853230 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.853261 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.853271 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.853286 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.853297 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.858435 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.867379 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.875066 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:53Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.890596 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/09bf1dd6-0f35-4ec3-9d94-87800241f5eb-hosts-file\") pod \"node-resolver-rf7dv\" (UID: \"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\") " pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.890651 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xx2m\" (UniqueName: \"kubernetes.io/projected/09bf1dd6-0f35-4ec3-9d94-87800241f5eb-kube-api-access-5xx2m\") pod \"node-resolver-rf7dv\" (UID: \"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\") " pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.890677 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-serviceca\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.890708 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-host\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.890764 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kt2d\" (UniqueName: \"kubernetes.io/projected/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-kube-api-access-2kt2d\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.891144 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/09bf1dd6-0f35-4ec3-9d94-87800241f5eb-hosts-file\") pod \"node-resolver-rf7dv\" (UID: \"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\") " pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.892386 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-host\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.892887 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-serviceca\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.911067 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xx2m\" (UniqueName: \"kubernetes.io/projected/09bf1dd6-0f35-4ec3-9d94-87800241f5eb-kube-api-access-5xx2m\") pod \"node-resolver-rf7dv\" (UID: \"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\") " pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.911620 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kt2d\" (UniqueName: \"kubernetes.io/projected/7f2ccf6c-bde8-4835-b4f5-5eaee2916835-kube-api-access-2kt2d\") pod \"node-ca-b26td\" (UID: \"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\") " pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.956105 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.956148 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.956158 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.956172 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:53 crc kubenswrapper[4838]: I1207 09:06:53.956183 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:53Z","lastTransitionTime":"2025-12-07T09:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.006807 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-b26td" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.017624 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rf7dv" Dec 07 09:06:54 crc kubenswrapper[4838]: W1207 09:06:54.020357 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f2ccf6c_bde8_4835_b4f5_5eaee2916835.slice/crio-7ab0ba3c4987549cd3d691c28b94aa4348405418dc93da92c53f22a7d9f32321 WatchSource:0}: Error finding container 7ab0ba3c4987549cd3d691c28b94aa4348405418dc93da92c53f22a7d9f32321: Status 404 returned error can't find the container with id 7ab0ba3c4987549cd3d691c28b94aa4348405418dc93da92c53f22a7d9f32321 Dec 07 09:06:54 crc kubenswrapper[4838]: W1207 09:06:54.030091 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09bf1dd6_0f35_4ec3_9d94_87800241f5eb.slice/crio-e6f7bc75d9bdd2476798b008b0c78497eab9cf1f1eadf90cd3c88d829d412b0b WatchSource:0}: Error finding container e6f7bc75d9bdd2476798b008b0c78497eab9cf1f1eadf90cd3c88d829d412b0b: Status 404 returned error can't find the container with id e6f7bc75d9bdd2476798b008b0c78497eab9cf1f1eadf90cd3c88d829d412b0b Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.060383 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.060415 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.060424 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.060437 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.060446 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.165937 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.165962 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.165971 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.165984 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.165992 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.167848 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7v9qs"] Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.168152 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.171708 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.171727 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.171781 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.171948 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.171997 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.193889 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.213276 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.223419 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.244409 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.262629 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.267752 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.267784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.267796 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.267810 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.267841 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.274898 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.291395 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.294905 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d1304517-3ac2-4aa3-afe2-01e27f87d880-rootfs\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.294936 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tphnq\" (UniqueName: \"kubernetes.io/projected/d1304517-3ac2-4aa3-afe2-01e27f87d880-kube-api-access-tphnq\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.294960 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1304517-3ac2-4aa3-afe2-01e27f87d880-mcd-auth-proxy-config\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.294978 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1304517-3ac2-4aa3-afe2-01e27f87d880-proxy-tls\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.311522 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.329633 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.347869 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.370362 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.370392 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.370402 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.370416 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.370424 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.396145 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1304517-3ac2-4aa3-afe2-01e27f87d880-mcd-auth-proxy-config\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.396180 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1304517-3ac2-4aa3-afe2-01e27f87d880-proxy-tls\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.396209 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tphnq\" (UniqueName: \"kubernetes.io/projected/d1304517-3ac2-4aa3-afe2-01e27f87d880-kube-api-access-tphnq\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.396229 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d1304517-3ac2-4aa3-afe2-01e27f87d880-rootfs\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.396281 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d1304517-3ac2-4aa3-afe2-01e27f87d880-rootfs\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.396890 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1304517-3ac2-4aa3-afe2-01e27f87d880-mcd-auth-proxy-config\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.399959 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1304517-3ac2-4aa3-afe2-01e27f87d880-proxy-tls\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.411773 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tphnq\" (UniqueName: \"kubernetes.io/projected/d1304517-3ac2-4aa3-afe2-01e27f87d880-kube-api-access-tphnq\") pod \"machine-config-daemon-7v9qs\" (UID: \"d1304517-3ac2-4aa3-afe2-01e27f87d880\") " pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.472976 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.473007 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.473016 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.473030 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.473038 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.491936 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:06:54 crc kubenswrapper[4838]: W1207 09:06:54.503361 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1304517_3ac2_4aa3_afe2_01e27f87d880.slice/crio-8a8e201c20b1e46c415dc9f6c353a20002f1b3b60eae456fd62ffffe1296ca28 WatchSource:0}: Error finding container 8a8e201c20b1e46c415dc9f6c353a20002f1b3b60eae456fd62ffffe1296ca28: Status 404 returned error can't find the container with id 8a8e201c20b1e46c415dc9f6c353a20002f1b3b60eae456fd62ffffe1296ca28 Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.535477 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-bgkt9"] Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.535762 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.537522 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-jkrnf"] Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.538073 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.539793 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.539844 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.540208 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.540444 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.540912 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.547875 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.548132 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.548145 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-brr6g"] Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.548939 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: W1207 09:06:54.552478 4838 reflector.go:561] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 07 09:06:54 crc kubenswrapper[4838]: E1207 09:06:54.552527 4838 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 07 09:06:54 crc kubenswrapper[4838]: W1207 09:06:54.552482 4838 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 07 09:06:54 crc kubenswrapper[4838]: E1207 09:06:54.552562 4838 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.564467 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.568863 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.568936 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.569144 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.576831 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.577769 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.577798 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.577807 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.577974 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.577986 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.581797 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.595477 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-07 09:01:53 +0000 UTC, rotation deadline is 2026-10-01 18:58:39.011302932 +0000 UTC Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.595531 4838 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7161h51m44.415774089s for next certificate rotation Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.601218 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.614164 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:54 crc kubenswrapper[4838]: E1207 09:06:54.614277 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.640045 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.679916 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.680282 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.680322 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.680331 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.680342 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.680351 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.698747 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.698840 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-hostroot\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.698877 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cni-binary-copy\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.698898 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwpxt\" (UniqueName: \"kubernetes.io/projected/5eb803b5-3546-42b8-829e-ba158fb6a2d3-kube-api-access-fwpxt\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.698926 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-var-lib-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.698942 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.698963 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-config\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699013 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66f9850-4704-46f3-9745-bdf80caf8a98-ovn-node-metrics-cert\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699062 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-cnibin\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699083 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-k8s-cni-cncf-io\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699101 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-netns\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699139 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-cni-multus\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699172 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-log-socket\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699188 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699224 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-bin\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699240 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-cni-binary-copy\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699274 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-daemon-config\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699298 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-socket-dir-parent\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699312 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-etc-kubernetes\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699371 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-os-release\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699391 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-ovn-kubernetes\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699432 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699449 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-env-overrides\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699464 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-slash\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699517 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-etc-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699535 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-conf-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699549 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-system-cni-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699599 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-cni-bin\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699614 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cnibin\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699633 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-systemd-units\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699696 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-kubelet\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699718 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-system-cni-dir\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699734 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-os-release\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699806 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-ovn\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699846 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-script-lib\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699863 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-netns\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699876 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-cni-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699889 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-multus-certs\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699903 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699934 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-kubelet\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699959 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-netd\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.699975 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-node-log\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.700000 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcrtj\" (UniqueName: \"kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.700017 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6ln5\" (UniqueName: \"kubernetes.io/projected/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-kube-api-access-g6ln5\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.700052 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-systemd\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.718880 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.733448 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.745475 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.755681 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.764650 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.774400 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.782495 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.782535 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.782547 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.782599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.782609 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.786590 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.797869 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800704 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800742 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-config\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800766 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66f9850-4704-46f3-9745-bdf80caf8a98-ovn-node-metrics-cert\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800788 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-cnibin\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800808 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-k8s-cni-cncf-io\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800847 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-netns\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800868 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-cni-multus\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800898 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-log-socket\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800918 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800937 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-daemon-config\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800956 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-bin\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800976 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-cni-binary-copy\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.800997 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-etc-kubernetes\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801028 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-socket-dir-parent\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801052 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-ovn-kubernetes\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801062 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-netns\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801071 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-os-release\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801137 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-env-overrides\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801141 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-os-release\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801156 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801175 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-slash\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801193 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-etc-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801213 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-conf-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801231 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-systemd-units\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801246 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-system-cni-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801262 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-cni-bin\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801278 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cnibin\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801295 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-ovn\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801315 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-kubelet\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801335 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-slash\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801369 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-system-cni-dir\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801383 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801400 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-cnibin\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801407 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-systemd-units\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801451 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-etc-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801192 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-cni-multus\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801490 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-system-cni-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801493 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-conf-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801216 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-log-socket\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801537 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-cni-bin\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801560 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cnibin\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801587 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-ovn\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801608 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-var-lib-kubelet\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801981 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-env-overrides\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802014 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-k8s-cni-cncf-io\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802050 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-socket-dir-parent\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802074 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-etc-kubernetes\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802098 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-ovn-kubernetes\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802111 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-config\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802120 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.801337 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-system-cni-dir\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802144 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-bin\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802156 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-daemon-config\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802162 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-os-release\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802188 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-netns\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802291 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-os-release\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802304 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5eb803b5-3546-42b8-829e-ba158fb6a2d3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802312 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-script-lib\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802322 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-netns\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802353 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-netd\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802393 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-netd\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802421 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-cni-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802433 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-cni-binary-copy\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802472 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-multus-cni-dir\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802478 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-multus-certs\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802508 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802533 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-kubelet\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802587 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-systemd\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802606 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-node-log\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802629 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcrtj\" (UniqueName: \"kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802633 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-systemd\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802653 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6ln5\" (UniqueName: \"kubernetes.io/projected/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-kube-api-access-g6ln5\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802670 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-kubelet\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802675 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-var-lib-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802608 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-host-run-multus-certs\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802697 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-hostroot\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802718 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-node-log\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802723 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cni-binary-copy\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802745 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwpxt\" (UniqueName: \"kubernetes.io/projected/5eb803b5-3546-42b8-829e-ba158fb6a2d3-kube-api-access-fwpxt\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802893 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-script-lib\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802951 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-hostroot\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.802991 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-var-lib-openvswitch\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.803037 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.803461 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5eb803b5-3546-42b8-829e-ba158fb6a2d3-cni-binary-copy\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.806523 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66f9850-4704-46f3-9745-bdf80caf8a98-ovn-node-metrics-cert\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.812289 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.820321 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rf7dv" event={"ID":"09bf1dd6-0f35-4ec3-9d94-87800241f5eb","Type":"ContainerStarted","Data":"dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.820346 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rf7dv" event={"ID":"09bf1dd6-0f35-4ec3-9d94-87800241f5eb","Type":"ContainerStarted","Data":"e6f7bc75d9bdd2476798b008b0c78497eab9cf1f1eadf90cd3c88d829d412b0b"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.821348 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-b26td" event={"ID":"7f2ccf6c-bde8-4835-b4f5-5eaee2916835","Type":"ContainerStarted","Data":"23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.821474 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-b26td" event={"ID":"7f2ccf6c-bde8-4835-b4f5-5eaee2916835","Type":"ContainerStarted","Data":"7ab0ba3c4987549cd3d691c28b94aa4348405418dc93da92c53f22a7d9f32321"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.823029 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.823059 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.823068 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"8a8e201c20b1e46c415dc9f6c353a20002f1b3b60eae456fd62ffffe1296ca28"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.826260 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwpxt\" (UniqueName: \"kubernetes.io/projected/5eb803b5-3546-42b8-829e-ba158fb6a2d3-kube-api-access-fwpxt\") pod \"multus-additional-cni-plugins-jkrnf\" (UID: \"5eb803b5-3546-42b8-829e-ba158fb6a2d3\") " pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.826269 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.830347 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6ln5\" (UniqueName: \"kubernetes.io/projected/ca7ddad3-e3c1-47c4-b4a8-d332f9d69227-kube-api-access-g6ln5\") pod \"multus-bgkt9\" (UID: \"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\") " pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.837170 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.847172 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.857001 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.860621 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bgkt9" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.868738 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.869176 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: W1207 09:06:54.877097 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca7ddad3_e3c1_47c4_b4a8_d332f9d69227.slice/crio-40a4f423b3bb2e9b1d90a4523338f29d1e4fb9d6677988c6f2e1a9857744c519 WatchSource:0}: Error finding container 40a4f423b3bb2e9b1d90a4523338f29d1e4fb9d6677988c6f2e1a9857744c519: Status 404 returned error can't find the container with id 40a4f423b3bb2e9b1d90a4523338f29d1e4fb9d6677988c6f2e1a9857744c519 Dec 07 09:06:54 crc kubenswrapper[4838]: W1207 09:06:54.880060 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5eb803b5_3546_42b8_829e_ba158fb6a2d3.slice/crio-fa09c38490b5797338caa538f597d88f844e0b0d7075865f2923089feede98ee WatchSource:0}: Error finding container fa09c38490b5797338caa538f597d88f844e0b0d7075865f2923089feede98ee: Status 404 returned error can't find the container with id fa09c38490b5797338caa538f597d88f844e0b0d7075865f2923089feede98ee Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.884612 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.884652 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.884665 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.884684 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.884697 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.895941 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.921087 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.933027 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.946056 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.958569 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.970669 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.985874 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.991206 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.991495 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.991505 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.991519 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.991530 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:54Z","lastTransitionTime":"2025-12-07T09:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:54 crc kubenswrapper[4838]: I1207 09:06:54.997911 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:54Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.009159 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.020119 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.030521 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.045161 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.058534 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.074374 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.084846 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.094429 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.094461 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.094471 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.094537 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.094577 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.099032 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.114854 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.130057 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.197038 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.197074 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.197085 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.197103 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.197113 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.298984 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.299014 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.299022 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.299036 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.299047 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.309271 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.309471 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:07:03.309453441 +0000 UTC m=+40.016772458 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.400996 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.401029 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.401041 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.401057 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.401068 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.410464 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.410509 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.410534 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.410579 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.410708 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.410735 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.410749 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.410799 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:03.410782573 +0000 UTC m=+40.118101600 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411138 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411185 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:03.411170284 +0000 UTC m=+40.118489311 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411248 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411263 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411274 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411304 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:03.411293888 +0000 UTC m=+40.118612915 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411350 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.411377 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:03.41136864 +0000 UTC m=+40.118687667 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.477590 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.504113 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.504140 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.504157 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.504169 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.504179 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.606631 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.606662 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.606669 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.606681 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.606690 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.615103 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.615196 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.615321 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.615422 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.708552 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.708586 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.708594 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.708607 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.708618 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.810726 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.810766 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.810775 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.810789 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.810798 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.816028 4838 projected.go:288] Couldn't get configMap openshift-ovn-kubernetes/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.816057 4838 projected.go:194] Error preparing data for projected volume kube-api-access-vcrtj for pod openshift-ovn-kubernetes/ovnkube-node-brr6g: failed to sync configmap cache: timed out waiting for the condition Dec 07 09:06:55 crc kubenswrapper[4838]: E1207 09:06:55.816118 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj podName:a66f9850-4704-46f3-9745-bdf80caf8a98 nodeName:}" failed. No retries permitted until 2025-12-07 09:06:56.316100391 +0000 UTC m=+33.023419398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vcrtj" (UniqueName: "kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj") pod "ovnkube-node-brr6g" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98") : failed to sync configmap cache: timed out waiting for the condition Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.829348 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eb803b5-3546-42b8-829e-ba158fb6a2d3" containerID="c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30" exitCode=0 Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.829399 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerDied","Data":"c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.829460 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerStarted","Data":"fa09c38490b5797338caa538f597d88f844e0b0d7075865f2923089feede98ee"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.831413 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerStarted","Data":"bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.831468 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerStarted","Data":"40a4f423b3bb2e9b1d90a4523338f29d1e4fb9d6677988c6f2e1a9857744c519"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.846963 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.863782 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.872658 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.884039 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.901306 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.918715 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.918747 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.918755 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.918769 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.918777 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:55Z","lastTransitionTime":"2025-12-07T09:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.919676 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.932593 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.944394 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.960396 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.974279 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.987250 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:55 crc kubenswrapper[4838]: I1207 09:06:55.997699 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:55Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.008375 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.020911 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.020964 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.020979 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.021000 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.021015 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.022458 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.034157 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.045632 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.059206 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.070519 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.079514 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.083993 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.101928 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.113941 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.123882 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.123921 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.123934 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.123953 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.123964 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.124955 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.135022 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.147127 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.159955 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.174915 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.225578 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.225610 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.225618 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.225630 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.225638 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.319456 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcrtj\" (UniqueName: \"kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.323851 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcrtj\" (UniqueName: \"kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj\") pod \"ovnkube-node-brr6g\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.332433 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.332467 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.332485 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.332501 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.332511 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.373782 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:06:56 crc kubenswrapper[4838]: W1207 09:06:56.386140 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda66f9850_4704_46f3_9745_bdf80caf8a98.slice/crio-3e8f79c96250a7a72924c64a1f8cec3f7c67c4ee40f4070770832d7870792fe3 WatchSource:0}: Error finding container 3e8f79c96250a7a72924c64a1f8cec3f7c67c4ee40f4070770832d7870792fe3: Status 404 returned error can't find the container with id 3e8f79c96250a7a72924c64a1f8cec3f7c67c4ee40f4070770832d7870792fe3 Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.434309 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.434334 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.434343 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.434355 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.434363 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.536911 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.536946 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.536956 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.536970 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.536980 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.614097 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:56 crc kubenswrapper[4838]: E1207 09:06:56.614259 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.639503 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.639538 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.639547 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.639561 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.639571 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.741887 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.741916 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.741924 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.741936 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.741945 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.836357 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eb803b5-3546-42b8-829e-ba158fb6a2d3" containerID="fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d" exitCode=0 Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.836409 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerDied","Data":"fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.838952 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636" exitCode=0 Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.838986 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.839018 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"3e8f79c96250a7a72924c64a1f8cec3f7c67c4ee40f4070770832d7870792fe3"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.843351 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.843376 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.843384 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.843395 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.843404 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.855378 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.870146 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.889047 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.908675 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.933225 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.945552 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.945580 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.945588 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.945599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.945607 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:56Z","lastTransitionTime":"2025-12-07T09:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.960795 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:56 crc kubenswrapper[4838]: I1207 09:06:56.995178 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:56Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.022452 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.047829 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.047856 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.047864 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.047876 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.047884 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.048989 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.064066 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.073539 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.087528 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.104560 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.115991 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.127634 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.137699 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.149741 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.149792 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.149805 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.149863 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.149878 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.151799 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.166206 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.179120 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.192200 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.205656 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.218140 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.229875 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.240973 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.249641 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.251745 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.251767 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.251776 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.251788 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.251796 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.266436 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.358730 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.358775 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.358784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.358799 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.358808 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.461295 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.461327 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.461335 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.461349 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.461357 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.563738 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.563777 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.563787 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.563801 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.563812 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.614091 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.614112 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:57 crc kubenswrapper[4838]: E1207 09:06:57.614213 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:06:57 crc kubenswrapper[4838]: E1207 09:06:57.614369 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.665836 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.665881 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.665895 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.665912 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.665925 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.768485 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.768536 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.768548 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.768565 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.768577 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.847399 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eb803b5-3546-42b8-829e-ba158fb6a2d3" containerID="181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610" exitCode=0 Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.847455 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerDied","Data":"181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.855251 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.855294 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.855305 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.855319 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.855329 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.855342 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.872269 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.872346 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.872370 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.872399 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.872419 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.877242 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.900502 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.919990 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.940519 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.955610 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.968290 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.975123 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.975161 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.975173 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.975190 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.975205 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:57Z","lastTransitionTime":"2025-12-07T09:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:57 crc kubenswrapper[4838]: I1207 09:06:57.980658 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:57.999945 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:57Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.018530 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.031020 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.041941 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.059137 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.077172 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.077241 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.077256 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.077322 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.077336 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.081368 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.180093 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.180137 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.180151 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.180173 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.180187 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.282912 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.282974 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.282991 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.283015 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.283034 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.386609 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.386660 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.386678 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.386700 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.386718 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.488997 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.489056 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.489075 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.489098 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.489117 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.574671 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.591808 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.591888 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.591909 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.591933 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.591952 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.597645 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.614413 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:06:58 crc kubenswrapper[4838]: E1207 09:06:58.614546 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.617662 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.638875 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.656972 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.675234 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.695132 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.695203 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.695229 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.695255 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.695410 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.697799 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.731235 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.749254 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.764776 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.780011 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.798701 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.798767 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.798784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.798808 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.798851 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.802547 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.818205 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.837088 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.862618 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eb803b5-3546-42b8-829e-ba158fb6a2d3" containerID="48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6" exitCode=0 Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.862683 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerDied","Data":"48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.892507 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.900902 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.900957 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.902109 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.902170 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.902190 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:58Z","lastTransitionTime":"2025-12-07T09:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.911943 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.925524 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.940975 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.968391 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:58 crc kubenswrapper[4838]: I1207 09:06:58.989034 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:58Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.004740 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.004805 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.004855 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.004874 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.004885 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.009431 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.024900 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.039971 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.053420 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.064878 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.077839 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.092974 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.106946 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.106975 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.106983 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.106998 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.107008 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.208894 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.208940 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.208958 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.208980 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.208996 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.312065 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.312317 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.312484 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.312628 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.312765 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.415043 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.415332 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.415451 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.415604 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.415727 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.525443 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.529674 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.529920 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.530079 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.530248 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.614529 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:06:59 crc kubenswrapper[4838]: E1207 09:06:59.614688 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.616761 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:06:59 crc kubenswrapper[4838]: E1207 09:06:59.617249 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.632343 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.632378 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.632391 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.632409 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.632422 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.735087 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.735136 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.735163 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.735187 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.735205 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.837515 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.837551 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.837559 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.837574 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.837583 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.871447 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.874089 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eb803b5-3546-42b8-829e-ba158fb6a2d3" containerID="5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d" exitCode=0 Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.874137 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerDied","Data":"5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.894908 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.914713 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.938033 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.942202 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.942282 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.942306 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.942735 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.942754 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:06:59Z","lastTransitionTime":"2025-12-07T09:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.954307 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.972549 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:06:59 crc kubenswrapper[4838]: I1207 09:06:59.988574 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:06:59.999944 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:06:59Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.020480 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.045856 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.045882 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.045890 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.045903 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.045912 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.050503 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.068301 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.088358 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.101275 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.114802 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.148311 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.148346 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.148357 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.148374 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.148385 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.250699 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.250759 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.250779 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.250805 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.250847 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.353087 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.353121 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.353132 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.353144 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.353152 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.455570 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.455599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.455609 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.455623 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.455633 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.558462 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.558528 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.558545 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.558569 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.558587 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.613913 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:00 crc kubenswrapper[4838]: E1207 09:07:00.614096 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.661313 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.661362 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.661377 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.661395 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.661409 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.764258 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.764329 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.764344 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.764365 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.764396 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.867630 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.868033 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.868072 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.868108 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.868149 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.884608 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eb803b5-3546-42b8-829e-ba158fb6a2d3" containerID="ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9" exitCode=0 Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.884676 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerDied","Data":"ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.901081 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.914554 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.931741 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.947554 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.961775 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.977493 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.981919 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.981944 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.981953 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.981967 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.981978 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:00Z","lastTransitionTime":"2025-12-07T09:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:00 crc kubenswrapper[4838]: I1207 09:07:00.991304 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:00Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.004240 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.020283 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.030866 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.040260 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.053228 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.074842 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.084454 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.084492 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.084502 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.084516 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.084528 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.186656 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.186683 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.186691 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.186704 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.186713 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.288834 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.288857 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.288866 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.288878 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.288887 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.391881 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.391932 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.391945 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.391964 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.391978 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.493691 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.493727 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.493736 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.493750 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.493761 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.596546 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.596593 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.596605 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.596620 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.596632 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.613510 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.613630 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:01 crc kubenswrapper[4838]: E1207 09:07:01.613793 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:01 crc kubenswrapper[4838]: E1207 09:07:01.614503 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.698753 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.698788 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.698798 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.698832 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.698846 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.801031 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.801058 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.801068 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.801080 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.801090 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.891118 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" event={"ID":"5eb803b5-3546-42b8-829e-ba158fb6a2d3","Type":"ContainerStarted","Data":"5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.896698 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.897122 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.902712 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.902754 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.902769 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.902788 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.902801 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:01Z","lastTransitionTime":"2025-12-07T09:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.909342 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.929195 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.942261 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.953062 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.969955 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.987630 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:01 crc kubenswrapper[4838]: I1207 09:07:01.989174 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:01Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.004608 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.004649 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.004659 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.004675 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.004688 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.009461 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.019913 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.029074 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.041751 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.059425 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.077358 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.099620 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.107109 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.107142 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.107153 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.107168 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.107179 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.117982 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.131061 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.143026 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.173927 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.194566 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.209203 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.209233 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.209241 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.209254 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.209263 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.223857 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.240719 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.249156 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.257663 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.265849 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.273176 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.285195 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.296846 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.311075 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.311101 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.311108 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.311122 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.311132 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.412809 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.412868 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.412878 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.412895 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.412905 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.515863 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.515927 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.515949 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.515978 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.516000 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.614056 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:02 crc kubenswrapper[4838]: E1207 09:07:02.614274 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.619620 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.619675 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.619696 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.619722 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.619739 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.723326 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.723394 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.723417 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.723444 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.723466 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.826005 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.826109 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.826139 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.826165 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.826182 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.901314 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.901386 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.928859 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.928910 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.928928 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.928951 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.928972 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:02Z","lastTransitionTime":"2025-12-07T09:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.937915 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.959257 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.973952 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:02 crc kubenswrapper[4838]: I1207 09:07:02.987367 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.001002 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.017261 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.032161 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.032204 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.032216 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.032232 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.032244 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.040872 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.058339 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.072721 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.090134 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.139683 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.139729 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.139743 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.139762 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.139776 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.146630 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.161840 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.186113 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.218254 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.242107 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.242152 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.242165 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.242183 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.242197 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.324875 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.324917 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.324927 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.324940 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.324949 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.338687 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.341695 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.341719 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.341727 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.341738 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.341747 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.352784 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.355900 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.355922 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.355930 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.355945 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.355953 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.366246 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.369045 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.369076 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.369085 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.369100 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.369110 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.382473 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.387545 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.387589 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.387606 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.387627 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.387642 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.400069 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.400280 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:07:19.400240721 +0000 UTC m=+56.107559738 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.400061 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.400417 4838 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.402286 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.402314 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.402325 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.402342 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.402353 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.480068 4838 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.503326 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.503378 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.503406 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.503430 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503534 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503556 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503574 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503632 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:19.503607192 +0000 UTC m=+56.210926209 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503729 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503755 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503769 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503654 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503853 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:19.503839908 +0000 UTC m=+56.211158925 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.503937 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:19.50391369 +0000 UTC m=+56.211232697 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.504010 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.504114 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:19.504084975 +0000 UTC m=+56.211403992 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.505256 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.505302 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.505315 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.505359 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.505371 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.608073 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.608690 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.608953 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.609200 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.609650 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.614308 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.614674 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.614976 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:03 crc kubenswrapper[4838]: E1207 09:07:03.615310 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.625663 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.641017 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.659138 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.672970 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.689370 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.701783 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.710864 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.712041 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.712064 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.712087 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.712099 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.712109 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.723429 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.735108 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.744746 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.755567 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.766539 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.785753 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.831322 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.831346 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.831371 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.831386 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.831395 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.933396 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.933433 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.933443 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.933459 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:03 crc kubenswrapper[4838]: I1207 09:07:03.933470 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:03Z","lastTransitionTime":"2025-12-07T09:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.035057 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.035315 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.035332 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.035348 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.035359 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.136774 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.136806 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.136830 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.136842 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.136851 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.239728 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.239771 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.239785 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.239804 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.239840 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.343035 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.343073 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.343083 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.343100 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.343112 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.446020 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.446068 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.446084 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.446105 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.446120 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.547762 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.547806 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.547844 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.547864 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.547875 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.613882 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:04 crc kubenswrapper[4838]: E1207 09:07:04.614098 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.650671 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.650703 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.650712 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.650724 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.650732 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.753395 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.753448 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.753464 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.753484 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.753498 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.855572 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.855627 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.855642 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.855661 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.855674 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.908730 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/0.log" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.912584 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7" exitCode=1 Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.912638 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.913849 4838 scope.go:117] "RemoveContainer" containerID="edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.931764 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.948925 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.958305 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.958351 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.958367 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.958389 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.958405 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:04Z","lastTransitionTime":"2025-12-07T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.964227 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.978521 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:04 crc kubenswrapper[4838]: I1207 09:07:04.994363 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.013615 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.028025 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.049552 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.061694 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.061773 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.061797 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.061856 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.061881 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.064109 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.079389 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.099615 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.113562 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.147366 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"\\\\nI1207 09:07:03.732040 6088 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732154 6088 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732324 6088 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732509 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 09:07:03.732586 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1207 09:07:03.732596 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1207 09:07:03.732611 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 09:07:03.732629 6088 factory.go:656] Stopping watch factory\\\\nI1207 09:07:03.732660 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 09:07:03.732519 6088 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1207 09:07:03.732859 6088 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 09:07:03.732867 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1207 09:07:03.732881 6088 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.169382 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.169416 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.169425 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.169439 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.169447 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.271325 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.271350 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.271358 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.271385 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.271394 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.373245 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.373275 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.373284 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.373297 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.373306 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.476589 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.476657 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.476679 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.476707 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.476728 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.579251 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.579292 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.579302 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.579314 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.579322 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.614368 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.614383 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:05 crc kubenswrapper[4838]: E1207 09:07:05.614531 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:05 crc kubenswrapper[4838]: E1207 09:07:05.614610 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.682201 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.682259 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.682277 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.682300 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.682316 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.785209 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.785243 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.785251 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.785264 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.785272 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.887790 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.887862 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.887878 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.887900 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.887915 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.917119 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/0.log" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.919632 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.920030 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.932135 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.943712 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.954470 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.964185 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.975630 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.989087 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.989700 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.989728 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.989738 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.989753 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.989764 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:05Z","lastTransitionTime":"2025-12-07T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:05 crc kubenswrapper[4838]: I1207 09:07:05.997292 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.008621 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.020450 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.030033 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.041419 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.051460 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.067906 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"\\\\nI1207 09:07:03.732040 6088 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732154 6088 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732324 6088 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732509 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 09:07:03.732586 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1207 09:07:03.732596 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1207 09:07:03.732611 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 09:07:03.732629 6088 factory.go:656] Stopping watch factory\\\\nI1207 09:07:03.732660 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 09:07:03.732519 6088 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1207 09:07:03.732859 6088 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 09:07:03.732867 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1207 09:07:03.732881 6088 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.091238 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.091262 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.091271 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.091284 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.091292 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.194601 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.194655 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.194673 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.194709 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.194728 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.297247 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.297299 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.297312 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.297333 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.297348 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.459016 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.459073 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.459093 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.459115 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.459133 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.561270 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.561362 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.561392 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.561426 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.561450 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.614353 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:06 crc kubenswrapper[4838]: E1207 09:07:06.614513 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.664460 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.664504 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.664517 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.664535 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.664548 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.767269 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.767302 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.767313 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.767328 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.767339 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.875979 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.876067 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.876084 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.876108 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.876126 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.906867 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb"] Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.907571 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.910200 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.910687 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.925633 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/1.log" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.926610 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/0.log" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.930133 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227" exitCode=1 Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.930181 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.930244 4838 scope.go:117] "RemoveContainer" containerID="edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.931577 4838 scope.go:117] "RemoveContainer" containerID="d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227" Dec 07 09:07:06 crc kubenswrapper[4838]: E1207 09:07:06.932128 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.937578 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.950594 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.966299 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.982538 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.982624 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.982648 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.982705 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.982718 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:06Z","lastTransitionTime":"2025-12-07T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:06 crc kubenswrapper[4838]: I1207 09:07:06.986202 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.002678 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.021320 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.040053 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.052044 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.064244 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9101534a-8a97-4d83-a97f-b8a0df78438d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.064536 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gljhm\" (UniqueName: \"kubernetes.io/projected/9101534a-8a97-4d83-a97f-b8a0df78438d-kube-api-access-gljhm\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.064705 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9101534a-8a97-4d83-a97f-b8a0df78438d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.064835 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9101534a-8a97-4d83-a97f-b8a0df78438d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.067591 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.084529 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"\\\\nI1207 09:07:03.732040 6088 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732154 6088 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732324 6088 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732509 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 09:07:03.732586 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1207 09:07:03.732596 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1207 09:07:03.732611 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 09:07:03.732629 6088 factory.go:656] Stopping watch factory\\\\nI1207 09:07:03.732660 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 09:07:03.732519 6088 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1207 09:07:03.732859 6088 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 09:07:03.732867 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1207 09:07:03.732881 6088 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.085680 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.085723 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.085732 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.085759 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.085770 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.095713 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.106054 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.116835 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.129625 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.143934 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.157422 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.166312 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9101534a-8a97-4d83-a97f-b8a0df78438d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.166362 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9101534a-8a97-4d83-a97f-b8a0df78438d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.166397 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9101534a-8a97-4d83-a97f-b8a0df78438d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.166437 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gljhm\" (UniqueName: \"kubernetes.io/projected/9101534a-8a97-4d83-a97f-b8a0df78438d-kube-api-access-gljhm\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.167354 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9101534a-8a97-4d83-a97f-b8a0df78438d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.168079 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9101534a-8a97-4d83-a97f-b8a0df78438d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.174496 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9101534a-8a97-4d83-a97f-b8a0df78438d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.180655 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.188906 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gljhm\" (UniqueName: \"kubernetes.io/projected/9101534a-8a97-4d83-a97f-b8a0df78438d-kube-api-access-gljhm\") pod \"ovnkube-control-plane-749d76644c-b77qb\" (UID: \"9101534a-8a97-4d83-a97f-b8a0df78438d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.189224 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.189261 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.189273 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.189290 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.189301 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.198874 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.210180 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.222345 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.224472 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.244166 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"\\\\nI1207 09:07:03.732040 6088 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732154 6088 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732324 6088 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732509 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 09:07:03.732586 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1207 09:07:03.732596 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1207 09:07:03.732611 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 09:07:03.732629 6088 factory.go:656] Stopping watch factory\\\\nI1207 09:07:03.732660 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 09:07:03.732519 6088 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1207 09:07:03.732859 6088 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 09:07:03.732867 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1207 09:07:03.732881 6088 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: W1207 09:07:07.248788 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9101534a_8a97_4d83_a97f_b8a0df78438d.slice/crio-536b25ca35d459f8acc86f8b726e656257a02af61b9b466b8e0c939baa8d506d WatchSource:0}: Error finding container 536b25ca35d459f8acc86f8b726e656257a02af61b9b466b8e0c939baa8d506d: Status 404 returned error can't find the container with id 536b25ca35d459f8acc86f8b726e656257a02af61b9b466b8e0c939baa8d506d Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.261323 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.284400 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.292307 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.292344 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.292355 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.292372 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.292387 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.295460 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.307518 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.322312 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.338405 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.354397 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.395079 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.395117 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.395132 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.395148 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.395157 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.497403 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.497443 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.497457 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.497473 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.497487 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.599864 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.599922 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.599938 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.599960 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.599974 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.614277 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.614382 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:07 crc kubenswrapper[4838]: E1207 09:07:07.614533 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:07 crc kubenswrapper[4838]: E1207 09:07:07.614656 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.702495 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.702542 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.702553 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.702573 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.702584 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.806074 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.806118 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.806128 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.806146 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.806158 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.910773 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.910841 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.910858 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.910881 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.910897 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:07Z","lastTransitionTime":"2025-12-07T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.941746 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" event={"ID":"9101534a-8a97-4d83-a97f-b8a0df78438d","Type":"ContainerStarted","Data":"57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.942094 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" event={"ID":"9101534a-8a97-4d83-a97f-b8a0df78438d","Type":"ContainerStarted","Data":"ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.942303 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" event={"ID":"9101534a-8a97-4d83-a97f-b8a0df78438d","Type":"ContainerStarted","Data":"536b25ca35d459f8acc86f8b726e656257a02af61b9b466b8e0c939baa8d506d"} Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.945802 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/1.log" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.952403 4838 scope.go:117] "RemoveContainer" containerID="d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227" Dec 07 09:07:07 crc kubenswrapper[4838]: E1207 09:07:07.952690 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.970025 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:07 crc kubenswrapper[4838]: I1207 09:07:07.995561 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.013492 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.013535 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.013551 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.013571 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.013588 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.016924 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.038211 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.056025 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.072009 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.086071 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.102682 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.114867 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.115878 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.115921 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.115936 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.115959 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.115976 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.130898 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edcebc1de5d1a7627e8804aaf4bd095d53166be04c11bb5722226478111b93c7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:03Z\\\",\\\"message\\\":\\\"\\\\nI1207 09:07:03.732040 6088 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732154 6088 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732324 6088 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 09:07:03.732509 6088 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 09:07:03.732586 6088 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1207 09:07:03.732596 6088 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1207 09:07:03.732611 6088 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 09:07:03.732629 6088 factory.go:656] Stopping watch factory\\\\nI1207 09:07:03.732660 6088 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 09:07:03.732519 6088 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1207 09:07:03.732859 6088 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 09:07:03.732867 6088 handler.go:208] Removed *v1.Node event handler 2\\\\nI1207 09:07:03.732881 6088 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.140899 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.151225 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.161305 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.169869 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.178653 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.186237 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.195627 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.207755 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.218170 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.218200 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.218208 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.218221 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.218229 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.223386 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.233835 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.253573 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.267468 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.281089 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.291256 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.304679 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.320768 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.320803 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.320828 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.320845 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.320855 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.320794 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.337310 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.350301 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.381759 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7nmsx"] Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.382545 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:08 crc kubenswrapper[4838]: E1207 09:07:08.382647 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.404177 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.423720 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.423777 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.423685 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.423795 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.423960 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.423980 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.441947 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.459423 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.480158 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c75kz\" (UniqueName: \"kubernetes.io/projected/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-kube-api-access-c75kz\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.480206 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.497450 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.525963 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.525996 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.526006 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.526018 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.526028 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.526611 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.536158 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.553573 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.565192 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.580831 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.581002 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c75kz\" (UniqueName: \"kubernetes.io/projected/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-kube-api-access-c75kz\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.581046 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:08 crc kubenswrapper[4838]: E1207 09:07:08.581162 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:08 crc kubenswrapper[4838]: E1207 09:07:08.581215 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:07:09.081201641 +0000 UTC m=+45.788520658 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.592225 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.598543 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c75kz\" (UniqueName: \"kubernetes.io/projected/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-kube-api-access-c75kz\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.610190 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.613853 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:08 crc kubenswrapper[4838]: E1207 09:07:08.613969 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.624171 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.628041 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.628069 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.628080 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.628094 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.628103 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.637642 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.648366 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.731136 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.731192 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.731207 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.731222 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.731232 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.833584 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.833642 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.833662 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.833686 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.833704 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.937080 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.937149 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.937171 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.937200 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:08 crc kubenswrapper[4838]: I1207 09:07:08.937222 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:08Z","lastTransitionTime":"2025-12-07T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.040220 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.040289 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.040313 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.040336 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.040353 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.087142 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:09 crc kubenswrapper[4838]: E1207 09:07:09.087350 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:09 crc kubenswrapper[4838]: E1207 09:07:09.087449 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:07:10.087416759 +0000 UTC m=+46.794735816 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.143204 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.143271 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.143295 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.143323 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.143342 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.246685 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.246774 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.246792 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.246859 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.246893 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.350407 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.350483 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.350496 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.350514 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.350525 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.453723 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.453778 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.453795 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.453843 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.453860 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.556596 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.556655 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.556675 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.556698 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.556717 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.614448 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:09 crc kubenswrapper[4838]: E1207 09:07:09.614980 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.615239 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.615246 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:09 crc kubenswrapper[4838]: E1207 09:07:09.615488 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:09 crc kubenswrapper[4838]: E1207 09:07:09.615543 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.660206 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.660256 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.660272 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.660292 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.660305 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.763244 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.763314 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.763339 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.763364 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.763383 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.866534 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.866622 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.866641 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.866666 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.866683 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.970046 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.970110 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.970134 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.970217 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:09 crc kubenswrapper[4838]: I1207 09:07:09.970241 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:09Z","lastTransitionTime":"2025-12-07T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.073264 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.073316 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.073332 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.073354 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.073371 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.096964 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:10 crc kubenswrapper[4838]: E1207 09:07:10.097182 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:10 crc kubenswrapper[4838]: E1207 09:07:10.097282 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:07:12.09725519 +0000 UTC m=+48.804574247 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.176795 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.176948 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.176971 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.177005 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.177027 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.280450 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.280532 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.280550 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.280573 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.280593 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.383604 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.383651 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.383669 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.383700 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.383717 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.486337 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.486391 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.486411 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.486432 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.486448 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.588915 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.588979 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.589002 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.589031 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.589051 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.613460 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:10 crc kubenswrapper[4838]: E1207 09:07:10.613611 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.692230 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.692287 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.692314 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.692342 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.692363 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.794715 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.794807 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.794855 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.794878 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.794898 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.898156 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.898224 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.898245 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.898272 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:10 crc kubenswrapper[4838]: I1207 09:07:10.898293 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:10Z","lastTransitionTime":"2025-12-07T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.001097 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.001179 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.001197 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.001221 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.001239 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.105015 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.105088 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.105111 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.105143 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.105164 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.208154 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.208218 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.208241 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.208270 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.208292 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.311306 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.311363 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.311380 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.311403 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.311419 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.414170 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.414226 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.414240 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.414261 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.414288 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.517399 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.517453 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.517464 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.517490 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.517503 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.614370 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.614413 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:11 crc kubenswrapper[4838]: E1207 09:07:11.614536 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.614594 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:11 crc kubenswrapper[4838]: E1207 09:07:11.614645 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:11 crc kubenswrapper[4838]: E1207 09:07:11.614770 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.620533 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.620565 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.620577 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.620595 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.620608 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.723492 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.723538 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.723554 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.723576 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.723592 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.830468 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.830527 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.830543 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.830566 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.830582 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.933083 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.933253 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.933278 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.933310 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:11 crc kubenswrapper[4838]: I1207 09:07:11.933335 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:11Z","lastTransitionTime":"2025-12-07T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.035899 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.035962 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.035981 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.036019 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.036039 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.117102 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:12 crc kubenswrapper[4838]: E1207 09:07:12.117369 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:12 crc kubenswrapper[4838]: E1207 09:07:12.117482 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:07:16.117455804 +0000 UTC m=+52.824774851 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.138753 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.138839 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.138857 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.138880 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.138897 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.241566 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.241640 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.241662 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.241694 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.241716 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.344796 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.344899 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.344963 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.344997 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.345020 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.447948 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.448054 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.448083 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.448113 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.448134 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.551409 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.551509 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.551566 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.551594 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.551611 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.613538 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:12 crc kubenswrapper[4838]: E1207 09:07:12.613761 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.655006 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.655068 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.655087 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.655108 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.655125 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.759243 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.759301 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.759326 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.759353 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.759374 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.862420 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.862475 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.862491 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.862512 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.862528 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.965293 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.965378 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.965409 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.965438 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:12 crc kubenswrapper[4838]: I1207 09:07:12.965458 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:12Z","lastTransitionTime":"2025-12-07T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.068303 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.068378 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.068403 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.068428 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.068448 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.172433 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.172534 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.172553 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.172579 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.172596 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.276082 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.276155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.276178 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.276212 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.276234 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.379152 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.379266 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.379301 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.379323 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.379343 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.482964 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.483416 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.483449 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.483479 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.483500 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.587009 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.587137 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.587158 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.587191 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.587229 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.613569 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.613630 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.613728 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.613778 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.613889 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.614103 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.637269 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.662005 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.680642 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.690089 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.690130 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.690144 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.690160 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.690172 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.700662 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.710166 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.710255 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.710284 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.710318 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.710352 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.726974 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.740603 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.744313 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.746674 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.746742 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.746766 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.746799 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.746857 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.765223 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.769009 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.774203 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.774257 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.774277 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.774301 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.774319 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.786107 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.795353 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.799435 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.799604 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.799629 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.799659 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.799677 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.803655 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.824224 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.829229 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.829318 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.829330 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.829348 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.829360 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.836364 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.850619 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: E1207 09:07:13.850870 4838 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.852739 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.852792 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.852845 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.852884 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.852909 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.857140 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.879788 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.903305 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.921339 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.938242 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.955314 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.955360 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.955371 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.955390 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:13 crc kubenswrapper[4838]: I1207 09:07:13.955402 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:13Z","lastTransitionTime":"2025-12-07T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.058230 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.058269 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.058280 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.058297 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.058308 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.161717 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.161902 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.161951 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.161978 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.162003 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.264555 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.264917 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.265112 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.265319 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.265472 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.367807 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.367929 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.367948 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.367980 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.368000 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.469784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.469852 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.469867 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.469883 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.469893 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.572884 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.572954 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.572974 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.572999 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.573016 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.614536 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:14 crc kubenswrapper[4838]: E1207 09:07:14.614676 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.675935 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.675985 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.675999 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.676018 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.676031 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.778300 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.778360 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.778377 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.778400 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.778417 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.880753 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.880801 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.880840 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.880861 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.880870 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.984187 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.984253 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.984273 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.984299 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:14 crc kubenswrapper[4838]: I1207 09:07:14.984325 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:14Z","lastTransitionTime":"2025-12-07T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.086264 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.086318 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.086335 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.086357 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.086375 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.188619 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.188653 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.188661 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.188674 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.188683 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.291758 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.291810 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.291886 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.291921 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.291944 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.394389 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.394479 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.394501 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.394530 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.394549 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.497684 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.497720 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.497730 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.497747 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.497758 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.600668 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.600734 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.600750 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.600771 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.600788 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.614053 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.614201 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.614053 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:15 crc kubenswrapper[4838]: E1207 09:07:15.614240 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:15 crc kubenswrapper[4838]: E1207 09:07:15.614367 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:15 crc kubenswrapper[4838]: E1207 09:07:15.614551 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.703261 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.703298 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.703309 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.703324 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.703335 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.807252 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.807689 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.807708 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.807731 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.807757 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.911211 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.911293 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.911312 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.911335 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.911355 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:15Z","lastTransitionTime":"2025-12-07T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.975942 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.994353 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 07 09:07:15 crc kubenswrapper[4838]: I1207 09:07:15.996920 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.014560 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.014610 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.014623 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.014640 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.014651 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.016887 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.036117 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.051676 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.071122 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.092915 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.107476 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.118002 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.118048 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.118063 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.118082 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.118095 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.121456 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.137976 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.154365 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.168179 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:16 crc kubenswrapper[4838]: E1207 09:07:16.168496 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:16 crc kubenswrapper[4838]: E1207 09:07:16.168601 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:07:24.168570024 +0000 UTC m=+60.875889091 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.181635 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.201106 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.217510 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.221230 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.221270 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.221285 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.221304 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.221319 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.250041 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.264289 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.324540 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.324600 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.324618 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.324645 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.324663 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.427081 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.427150 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.427174 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.427206 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.427230 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.530020 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.530097 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.530120 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.530152 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.530176 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.614134 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:16 crc kubenswrapper[4838]: E1207 09:07:16.614330 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.633106 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.633179 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.633204 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.633234 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.633258 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.735615 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.735683 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.735700 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.735723 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.735738 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.839069 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.839144 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.839155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.839179 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.839195 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.942207 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.942292 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.942319 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.942354 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:16 crc kubenswrapper[4838]: I1207 09:07:16.942375 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:16Z","lastTransitionTime":"2025-12-07T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.045343 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.045405 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.045425 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.045454 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.045480 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.148571 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.148631 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.148652 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.148676 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.148696 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.251704 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.251772 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.251794 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.251855 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.251883 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.354963 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.355032 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.355050 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.355076 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.355093 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.457694 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.458116 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.458382 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.458539 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.458655 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.561935 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.561988 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.562004 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.562027 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.562043 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.613605 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.613668 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.613615 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:17 crc kubenswrapper[4838]: E1207 09:07:17.613852 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:17 crc kubenswrapper[4838]: E1207 09:07:17.613923 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:17 crc kubenswrapper[4838]: E1207 09:07:17.614060 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.664993 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.665056 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.665073 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.665098 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.665115 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.767575 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.767645 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.767666 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.767689 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.767705 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.872267 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.872336 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.872354 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.872376 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.872392 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.976304 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.976360 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.976380 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.976404 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:17 crc kubenswrapper[4838]: I1207 09:07:17.976420 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:17Z","lastTransitionTime":"2025-12-07T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.079123 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.079155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.079164 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.079179 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.079189 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.182172 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.182229 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.182247 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.182270 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.182287 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.291599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.291632 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.291644 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.291660 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.291671 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.395230 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.395297 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.395316 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.395343 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.395362 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.498276 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.498339 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.498364 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.498393 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.498418 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.601005 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.601045 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.601056 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.601072 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.601084 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.613867 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:18 crc kubenswrapper[4838]: E1207 09:07:18.614056 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.615480 4838 scope.go:117] "RemoveContainer" containerID="d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.704599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.704701 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.704764 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.704793 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.704894 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.807395 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.807456 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.807474 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.807499 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.807519 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.909291 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.909330 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.909347 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.909372 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.909389 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:18Z","lastTransitionTime":"2025-12-07T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.994090 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/1.log" Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.996730 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7"} Dec 07 09:07:18 crc kubenswrapper[4838]: I1207 09:07:18.997204 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.011201 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.011410 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.011425 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.011434 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.011446 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.011455 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.023288 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.036679 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.051427 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.069967 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.090229 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.114081 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.114119 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.114129 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.114148 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.114160 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.119024 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.134388 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.151493 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.170395 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.181117 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.189705 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.198211 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.210557 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.215798 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.215837 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.215845 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.215857 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.215867 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.220213 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.254184 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.317755 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.317784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.317793 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.317804 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.317833 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.400737 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.401140 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:07:51.401098849 +0000 UTC m=+88.108417886 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.420760 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.420842 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.420858 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.420879 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.420894 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.522961 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.523000 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.523011 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.523026 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.523039 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.603644 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.603685 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.603704 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.603781 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:51.603759213 +0000 UTC m=+88.311078270 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.603469 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.604548 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.604618 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:51.604601347 +0000 UTC m=+88.311920404 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.604425 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.604701 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.604747 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.604928 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.604984 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.605033 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.605050 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.604999 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:51.604975678 +0000 UTC m=+88.312294735 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.605210 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:07:51.605188144 +0000 UTC m=+88.312507281 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.614421 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.614544 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.614620 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.614679 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.614851 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:19 crc kubenswrapper[4838]: E1207 09:07:19.614888 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.632397 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.632471 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.632507 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.632535 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.632554 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.739347 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.739411 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.739436 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.739463 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.739482 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.842451 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.842516 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.842532 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.842553 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.842567 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.946073 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.946148 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.946172 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.946199 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:19 crc kubenswrapper[4838]: I1207 09:07:19.946221 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:19Z","lastTransitionTime":"2025-12-07T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.003170 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/2.log" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.004262 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/1.log" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.007766 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7" exitCode=1 Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.007803 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.007867 4838 scope.go:117] "RemoveContainer" containerID="d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.008518 4838 scope.go:117] "RemoveContainer" containerID="3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7" Dec 07 09:07:20 crc kubenswrapper[4838]: E1207 09:07:20.008789 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.031227 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.052506 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.052595 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.052612 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.052660 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.052678 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.054554 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.072316 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.088026 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.102698 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.115468 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.135054 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d844241b3298035d04003423e5b1ca33d756b44ab49d8152b83a4635ea3c6227\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1207 09:07:05.702336 6204 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.145804 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.155718 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.155757 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.155773 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.155792 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.155807 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.159564 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.177961 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.192778 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.203422 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.214896 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.228176 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.243925 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.256237 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:20Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.257847 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.257961 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.258021 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.258101 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.258173 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.367909 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.367956 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.367970 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.367990 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.368005 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.471690 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.471759 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.471783 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.471847 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.471876 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.574934 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.574999 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.575022 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.575051 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.575074 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.613621 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:20 crc kubenswrapper[4838]: E1207 09:07:20.613844 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.677887 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.678295 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.678560 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.678774 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.678991 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.782119 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.782169 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.782185 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.782209 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.782225 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.884897 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.885241 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.885442 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.885651 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.885873 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.989249 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.989310 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.989334 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.989362 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:20 crc kubenswrapper[4838]: I1207 09:07:20.989384 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:20Z","lastTransitionTime":"2025-12-07T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.014322 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/2.log" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.020807 4838 scope.go:117] "RemoveContainer" containerID="3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7" Dec 07 09:07:21 crc kubenswrapper[4838]: E1207 09:07:21.021257 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.039969 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.060162 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.085083 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.092851 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.093130 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.093292 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.093452 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.093578 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.102097 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.119935 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.139773 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.163595 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.180760 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.194590 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.196323 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.196362 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.196375 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.196394 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.196408 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.209346 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.224721 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.238150 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.251480 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.262800 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.282436 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.293429 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.299078 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.299121 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.299136 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.299155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.299169 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.402212 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.402250 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.402261 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.402277 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.402288 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.505310 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.505353 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.505368 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.505387 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.505400 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.608024 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.608077 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.608093 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.608139 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.608157 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.613558 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.613614 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.613665 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:21 crc kubenswrapper[4838]: E1207 09:07:21.613881 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:21 crc kubenswrapper[4838]: E1207 09:07:21.614109 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:21 crc kubenswrapper[4838]: E1207 09:07:21.614433 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.711614 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.711674 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.711693 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.711718 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.711737 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.813681 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.813724 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.813735 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.813751 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.813765 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.916537 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.916613 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.916642 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.916674 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:21 crc kubenswrapper[4838]: I1207 09:07:21.916694 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:21Z","lastTransitionTime":"2025-12-07T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.018992 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.019040 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.019053 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.019071 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.019084 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.122209 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.122269 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.122286 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.122309 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.122329 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.224959 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.225017 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.225032 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.225052 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.225066 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.328049 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.328131 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.328148 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.328166 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.328178 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.431397 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.431465 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.431485 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.431508 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.431525 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.533471 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.533541 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.533558 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.533583 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.533611 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.614103 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:22 crc kubenswrapper[4838]: E1207 09:07:22.614280 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.636922 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.636980 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.637002 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.637032 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.637082 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.739949 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.740026 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.740050 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.740081 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.740105 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.848229 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.848264 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.848274 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.848462 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.848473 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.951730 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.951807 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.951872 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.951901 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:22 crc kubenswrapper[4838]: I1207 09:07:22.951922 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:22Z","lastTransitionTime":"2025-12-07T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.055207 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.055264 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.055283 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.055309 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.055326 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.158391 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.158468 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.158488 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.158512 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.158531 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.261645 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.261750 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.261773 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.261801 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.261855 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.366025 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.366409 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.366436 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.366467 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.366489 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.470992 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.471061 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.471083 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.471115 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.471136 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.573593 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.573715 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.573740 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.573772 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.573798 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.613516 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:23 crc kubenswrapper[4838]: E1207 09:07:23.613677 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.613744 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:23 crc kubenswrapper[4838]: E1207 09:07:23.613869 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.613949 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:23 crc kubenswrapper[4838]: E1207 09:07:23.614028 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.636601 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.655341 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.672073 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.676363 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.676400 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.676411 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.676429 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.676446 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.719412 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.742755 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.763597 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.779323 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.779359 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.779372 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.779390 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.779403 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.780369 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.792213 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.803083 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.815802 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.827576 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.839366 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.851216 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.862465 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.877946 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.881349 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.881502 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.881587 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.881668 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.881731 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.889182 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:23Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.984963 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.985037 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.985063 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.985093 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:23 crc kubenswrapper[4838]: I1207 09:07:23.985116 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:23Z","lastTransitionTime":"2025-12-07T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.087986 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.088042 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.088059 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.088087 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.088108 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.186569 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.187228 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.187555 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.187962 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.188362 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.210345 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.214765 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.214829 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.214845 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.214866 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.214877 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.232077 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.236360 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.236434 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.236457 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.236488 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.236510 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.252252 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.252533 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.252692 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:07:40.252657131 +0000 UTC m=+76.959976198 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.255342 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.260330 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.260391 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.260414 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.260445 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.260466 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.285031 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.292491 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.292561 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.292580 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.292604 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.292622 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.318153 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.318471 4838 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.321948 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.321994 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.322013 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.322035 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.322052 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.424433 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.424483 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.424496 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.424513 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.424525 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.526584 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.526623 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.526637 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.526656 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.526668 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.613903 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:24 crc kubenswrapper[4838]: E1207 09:07:24.614050 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.630941 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.630979 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.630992 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.631010 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.631023 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.733229 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.733264 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.733276 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.733292 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.733304 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.835428 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.835469 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.835483 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.835505 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.835518 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.937529 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.937626 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.937649 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.937679 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:24 crc kubenswrapper[4838]: I1207 09:07:24.937699 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:24Z","lastTransitionTime":"2025-12-07T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.040972 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.041036 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.041056 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.041081 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.041099 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.143899 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.143941 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.143952 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.143968 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.143979 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.247213 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.247625 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.247870 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.248092 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.248278 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.350929 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.350964 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.350976 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.350992 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.351006 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.453674 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.453705 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.453714 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.453726 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.453734 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.555682 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.555728 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.555745 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.555766 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.555784 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.613853 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.613905 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:25 crc kubenswrapper[4838]: E1207 09:07:25.613978 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.613853 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:25 crc kubenswrapper[4838]: E1207 09:07:25.614097 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:25 crc kubenswrapper[4838]: E1207 09:07:25.614143 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.658755 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.658784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.658793 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.658806 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.658833 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.762567 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.762675 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.762692 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.762715 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.762733 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.865379 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.865414 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.865425 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.865440 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.865451 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.968973 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.969039 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.969186 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.969583 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:25 crc kubenswrapper[4838]: I1207 09:07:25.969637 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:25Z","lastTransitionTime":"2025-12-07T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.071798 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.071881 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.071897 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.071916 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.071931 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.174606 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.174648 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.174658 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.174672 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.174680 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.277185 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.277421 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.277508 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.277633 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.277726 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.381441 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.381714 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.381727 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.381741 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.381750 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.484529 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.484574 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.484586 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.484607 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.484619 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.586652 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.586690 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.586706 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.586721 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.586732 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.614020 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:26 crc kubenswrapper[4838]: E1207 09:07:26.614205 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.688938 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.688987 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.688999 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.689017 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.689028 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.791563 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.791614 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.791625 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.791642 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.791655 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.893911 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.893960 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.893975 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.893993 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.894007 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.997529 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.997653 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.997679 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.997706 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:26 crc kubenswrapper[4838]: I1207 09:07:26.997726 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:26Z","lastTransitionTime":"2025-12-07T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.099990 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.100253 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.100378 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.100512 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.100634 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.203174 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.203219 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.203227 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.203241 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.203249 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.305766 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.306024 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.306116 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.306199 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.306257 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.408095 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.408421 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.408493 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.408557 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.408623 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.511336 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.511374 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.511384 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.511396 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.511405 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613342 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613378 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613387 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613402 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613412 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613886 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613890 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.613938 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:27 crc kubenswrapper[4838]: E1207 09:07:27.614203 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:27 crc kubenswrapper[4838]: E1207 09:07:27.614089 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:27 crc kubenswrapper[4838]: E1207 09:07:27.614300 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.715218 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.715263 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.715283 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.715299 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.715310 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.817608 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.817653 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.817662 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.817678 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.817697 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.920483 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.920513 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.920522 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.920535 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:27 crc kubenswrapper[4838]: I1207 09:07:27.920544 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:27Z","lastTransitionTime":"2025-12-07T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.023049 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.023079 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.023103 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.023117 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.023126 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.125537 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.125606 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.125623 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.125647 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.125669 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.227902 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.227955 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.227966 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.227981 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.227991 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.330801 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.330863 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.330876 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.330894 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.330906 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.433374 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.433415 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.433425 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.433442 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.433454 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.535810 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.535895 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.535981 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.536005 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.536021 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.613576 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:28 crc kubenswrapper[4838]: E1207 09:07:28.613695 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.638318 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.638361 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.638374 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.638390 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.638400 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.740733 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.740805 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.740838 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.740856 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.740869 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.842788 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.842861 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.842873 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.842888 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.842898 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.945142 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.945178 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.945187 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.945201 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:28 crc kubenswrapper[4838]: I1207 09:07:28.945216 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:28Z","lastTransitionTime":"2025-12-07T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.052654 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.052694 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.052704 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.052717 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.052726 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.155529 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.155579 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.155591 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.155609 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.155621 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.258553 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.258633 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.258657 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.258686 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.258710 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.360523 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.360615 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.360643 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.360671 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.360689 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.463310 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.463349 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.463359 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.463373 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.463381 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.565929 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.565982 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.565993 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.566008 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.566016 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.613722 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.613833 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.613740 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:29 crc kubenswrapper[4838]: E1207 09:07:29.613935 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:29 crc kubenswrapper[4838]: E1207 09:07:29.614020 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:29 crc kubenswrapper[4838]: E1207 09:07:29.614125 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.668554 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.668983 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.669371 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.670372 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.670537 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.773734 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.773776 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.773784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.773798 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.773807 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.876168 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.876247 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.876257 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.876272 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.876284 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.978847 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.978886 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.978907 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.978923 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:29 crc kubenswrapper[4838]: I1207 09:07:29.978935 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:29Z","lastTransitionTime":"2025-12-07T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.081465 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.081506 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.081515 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.081531 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.081539 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.183702 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.183763 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.183776 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.183789 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.183798 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.286639 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.286683 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.286695 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.286714 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.286726 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.388454 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.388487 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.388519 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.388532 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.388541 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.492250 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.492328 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.492354 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.492384 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.492407 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.595250 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.595493 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.595637 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.595736 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.595845 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.613987 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:30 crc kubenswrapper[4838]: E1207 09:07:30.614263 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.698848 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.698912 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.698936 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.698958 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.698972 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.801029 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.801071 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.801081 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.801096 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.801107 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.903884 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.904164 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.904330 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.904477 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:30 crc kubenswrapper[4838]: I1207 09:07:30.904646 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:30Z","lastTransitionTime":"2025-12-07T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.006458 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.006515 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.006532 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.006556 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.006572 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.108777 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.108862 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.108874 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.108891 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.108902 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.210639 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.210680 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.210690 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.210704 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.210713 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.313265 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.313532 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.313601 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.313671 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.313733 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.415676 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.416143 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.416300 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.416439 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.416556 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.519042 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.519072 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.519082 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.519094 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.519103 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.614412 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.614449 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:31 crc kubenswrapper[4838]: E1207 09:07:31.614630 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.614675 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:31 crc kubenswrapper[4838]: E1207 09:07:31.614754 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:31 crc kubenswrapper[4838]: E1207 09:07:31.614805 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.622546 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.622619 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.622644 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.622672 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.622705 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.725418 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.725455 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.725463 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.725478 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.725487 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.827911 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.827953 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.827965 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.827982 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.827996 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.930645 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.930688 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.930697 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.930711 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:31 crc kubenswrapper[4838]: I1207 09:07:31.930720 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:31Z","lastTransitionTime":"2025-12-07T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.032689 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.032736 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.032744 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.032759 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.032768 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.134681 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.134734 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.134746 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.134765 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.134777 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.237035 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.237075 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.237085 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.237101 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.237112 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.339732 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.339776 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.339786 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.339803 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.339858 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.442667 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.442716 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.442728 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.442744 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.442758 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.545530 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.545609 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.545621 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.545637 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.545646 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.614465 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:32 crc kubenswrapper[4838]: E1207 09:07:32.614646 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.647895 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.647932 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.647946 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.647967 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.647982 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.751172 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.751255 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.751281 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.751315 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.751342 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.853554 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.853601 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.853612 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.853628 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.853640 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.956339 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.956378 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.956387 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.956405 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:32 crc kubenswrapper[4838]: I1207 09:07:32.956416 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:32Z","lastTransitionTime":"2025-12-07T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.058245 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.058281 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.058291 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.058305 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.058313 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.160381 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.160419 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.160427 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.160441 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.160450 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.263883 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.263929 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.263940 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.263958 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.263970 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.366034 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.366087 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.366099 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.366115 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.366131 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.469024 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.469062 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.469077 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.469093 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.469102 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.571585 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.571631 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.571640 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.571654 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.571664 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.614110 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.614168 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.614110 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:33 crc kubenswrapper[4838]: E1207 09:07:33.614390 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:33 crc kubenswrapper[4838]: E1207 09:07:33.614527 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:33 crc kubenswrapper[4838]: E1207 09:07:33.614657 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.615885 4838 scope.go:117] "RemoveContainer" containerID="3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7" Dec 07 09:07:33 crc kubenswrapper[4838]: E1207 09:07:33.616399 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.628622 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.642065 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.657497 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.669993 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.673381 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.673434 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.673446 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.673463 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.673475 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.683087 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.693234 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.709540 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.721147 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.733554 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.749331 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.758011 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.764954 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.773890 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.775269 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.775299 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.775309 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.775325 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.775335 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.784160 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.794601 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.805642 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:33Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.877444 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.877491 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.877507 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.877528 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.877540 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.978980 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.979019 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.979029 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.979044 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:33 crc kubenswrapper[4838]: I1207 09:07:33.979057 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:33Z","lastTransitionTime":"2025-12-07T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.080981 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.081015 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.081028 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.081043 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.081053 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.183544 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.183576 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.183586 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.183602 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.183613 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.285415 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.285441 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.285450 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.285463 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.285472 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.388053 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.388083 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.388094 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.388107 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.388117 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.489953 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.490051 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.490072 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.490098 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.490118 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.525021 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.525063 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.525076 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.525091 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.525104 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: E1207 09:07:34.540106 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:34Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.544828 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.544952 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.545009 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.545090 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.545164 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: E1207 09:07:34.562661 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:34Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.565956 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.566016 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.566046 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.566064 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.566075 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: E1207 09:07:34.576691 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:34Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.579713 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.579744 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.579752 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.579765 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.579773 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: E1207 09:07:34.590295 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:34Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.592548 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.592572 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.592579 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.592590 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.592599 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: E1207 09:07:34.602543 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:34Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:34 crc kubenswrapper[4838]: E1207 09:07:34.602647 4838 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.604307 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.604331 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.604339 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.604354 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.604363 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.613504 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:34 crc kubenswrapper[4838]: E1207 09:07:34.613627 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.626111 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.706778 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.706804 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.706828 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.706840 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.706849 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.808998 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.809048 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.809058 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.809074 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.809083 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.911756 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.911810 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.911853 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.911876 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:34 crc kubenswrapper[4838]: I1207 09:07:34.911896 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:34Z","lastTransitionTime":"2025-12-07T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.014763 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.014800 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.014810 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.014839 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.014850 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.117001 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.117056 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.117067 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.117085 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.117098 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.219058 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.219106 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.219118 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.219136 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.219148 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.321415 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.321447 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.321455 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.321468 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.321476 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.424515 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.424552 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.424577 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.424594 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.424603 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.527336 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.527484 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.527511 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.527535 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.527552 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.614353 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.614360 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.614540 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:35 crc kubenswrapper[4838]: E1207 09:07:35.614701 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:35 crc kubenswrapper[4838]: E1207 09:07:35.614855 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:35 crc kubenswrapper[4838]: E1207 09:07:35.615061 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.630099 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.630149 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.630170 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.630194 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.630212 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.732881 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.732925 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.732937 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.732950 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.732959 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.836395 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.836447 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.836460 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.836478 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.836490 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.938960 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.939010 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.939022 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.939039 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:35 crc kubenswrapper[4838]: I1207 09:07:35.939050 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:35Z","lastTransitionTime":"2025-12-07T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.041654 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.041789 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.041804 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.041835 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.041848 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.144292 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.144327 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.144335 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.144350 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.144359 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.247190 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.247240 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.247249 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.247262 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.247271 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.349407 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.349446 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.349457 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.349471 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.349483 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.452032 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.452084 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.452101 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.452123 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.452140 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.555330 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.555368 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.555377 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.555392 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.555402 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.614574 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:36 crc kubenswrapper[4838]: E1207 09:07:36.614754 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.658302 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.658343 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.658351 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.658367 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.658379 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.761178 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.761214 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.761222 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.761235 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.761246 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.864122 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.864190 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.864214 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.864241 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.864262 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.967904 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.967969 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.967987 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.968014 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:36 crc kubenswrapper[4838]: I1207 09:07:36.968085 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:36Z","lastTransitionTime":"2025-12-07T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.071339 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.071409 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.071433 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.071461 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.071481 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.174231 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.174299 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.174321 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.174349 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.174370 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.278036 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.278111 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.278151 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.278181 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.278387 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.381805 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.381906 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.381927 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.381949 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.381967 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.485365 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.485414 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.485432 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.485456 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.485474 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.588851 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.588910 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.588924 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.588947 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.588961 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.614078 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.614224 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:37 crc kubenswrapper[4838]: E1207 09:07:37.614534 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.614759 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:37 crc kubenswrapper[4838]: E1207 09:07:37.615006 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:37 crc kubenswrapper[4838]: E1207 09:07:37.615185 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.690935 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.690989 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.691005 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.691025 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.691039 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.794199 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.794265 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.794291 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.794320 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.794341 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.897091 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.897143 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.897159 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.897181 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:37 crc kubenswrapper[4838]: I1207 09:07:37.897198 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:37Z","lastTransitionTime":"2025-12-07T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.000357 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.000420 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.000439 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.000461 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.000477 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.103656 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.103723 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.103740 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.103766 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.103784 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.207243 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.207315 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.207337 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.207360 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.207376 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.310140 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.310209 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.310227 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.310254 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.310273 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.413518 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.413572 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.413590 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.413614 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.413631 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.516720 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.516802 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.516855 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.516883 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.516901 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.614459 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:38 crc kubenswrapper[4838]: E1207 09:07:38.614717 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.620082 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.620146 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.620173 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.620213 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.620233 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.723757 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.723863 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.723889 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.723916 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.723934 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.826688 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.826759 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.826777 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.826801 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.826846 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.929806 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.929915 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.929935 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.929962 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:38 crc kubenswrapper[4838]: I1207 09:07:38.929980 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:38Z","lastTransitionTime":"2025-12-07T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.032691 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.032761 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.032801 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.032876 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.032907 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.135947 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.136007 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.136028 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.136054 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.136074 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.239115 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.239181 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.239203 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.239234 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.239256 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.342248 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.342347 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.342384 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.342425 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.342449 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.445341 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.445420 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.445445 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.445479 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.445501 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.548767 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.548863 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.548886 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.548915 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.548934 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.614155 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:39 crc kubenswrapper[4838]: E1207 09:07:39.614362 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.614185 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.614412 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:39 crc kubenswrapper[4838]: E1207 09:07:39.614567 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:39 crc kubenswrapper[4838]: E1207 09:07:39.614674 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.652688 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.652934 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.653016 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.653101 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.653133 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.756782 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.756919 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.756938 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.756964 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.756989 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.859897 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.859946 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.859983 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.860009 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.860027 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.962008 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.962050 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.962061 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.962079 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:39 crc kubenswrapper[4838]: I1207 09:07:39.962092 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:39Z","lastTransitionTime":"2025-12-07T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.064633 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.064713 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.064722 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.064736 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.064745 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.167664 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.167703 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.167714 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.167747 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.167756 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.271278 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.271358 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.271369 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.271386 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.271396 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.319740 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:40 crc kubenswrapper[4838]: E1207 09:07:40.320004 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:40 crc kubenswrapper[4838]: E1207 09:07:40.320186 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:08:12.320092265 +0000 UTC m=+109.027411363 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.374604 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.374679 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.374703 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.374731 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.374755 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.480245 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.480316 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.480338 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.480363 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.480377 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.583485 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.583522 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.583530 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.583544 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.583554 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.613682 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:40 crc kubenswrapper[4838]: E1207 09:07:40.613906 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.686214 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.686278 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.686301 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.686331 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.686353 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.790059 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.790139 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.790166 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.790199 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.790235 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.893250 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.893324 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.893349 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.893379 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.893403 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.997062 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.997139 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.997164 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.997193 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:40 crc kubenswrapper[4838]: I1207 09:07:40.997213 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:40Z","lastTransitionTime":"2025-12-07T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.080589 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/0.log" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.080666 4838 generic.go:334] "Generic (PLEG): container finished" podID="ca7ddad3-e3c1-47c4-b4a8-d332f9d69227" containerID="bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b" exitCode=1 Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.080727 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerDied","Data":"bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.081548 4838 scope.go:117] "RemoveContainer" containerID="bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.099097 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.099804 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.099897 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.099920 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.100999 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.101112 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.114943 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.137465 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.149506 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.162372 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.177380 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.196584 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.204169 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.204215 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.204232 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.204254 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.204272 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.210670 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.225606 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:40Z\\\",\\\"message\\\":\\\"2025-12-07T09:06:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73\\\\n2025-12-07T09:06:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73 to /host/opt/cni/bin/\\\\n2025-12-07T09:06:55Z [verbose] multus-daemon started\\\\n2025-12-07T09:06:55Z [verbose] Readiness Indicator file check\\\\n2025-12-07T09:07:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.240430 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cba62b5c-dd17-4a93-922d-237e5a7557bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://286fc6c21ddb56cdcb4bdb001dbbbcbd294b23971933e39148c05efed583e5b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.255966 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.270552 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.285497 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.299710 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.306544 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.306580 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.306592 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.306608 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.306621 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.313289 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.335243 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.348352 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.408926 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.408982 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.408995 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.409014 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.409026 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.511012 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.511092 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.511112 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.511137 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.511157 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.613485 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.613569 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.613485 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:41 crc kubenswrapper[4838]: E1207 09:07:41.613702 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:41 crc kubenswrapper[4838]: E1207 09:07:41.613806 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.613930 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.613983 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.614006 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.614034 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.614056 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: E1207 09:07:41.614127 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.715891 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.715949 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.715965 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.715994 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.716017 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.819182 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.819254 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.819278 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.819308 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.819332 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.922270 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.922312 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.922326 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.922343 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:41 crc kubenswrapper[4838]: I1207 09:07:41.922355 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:41Z","lastTransitionTime":"2025-12-07T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.024805 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.024938 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.024964 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.024998 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.025019 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.085666 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/0.log" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.085757 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerStarted","Data":"6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.106294 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.129734 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.129781 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.129794 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.129812 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.129855 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.130460 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.149373 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.164977 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.181643 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.200540 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.219451 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.232944 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.233000 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.233017 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.233041 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.233057 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.237454 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.255463 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.269792 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.289412 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:40Z\\\",\\\"message\\\":\\\"2025-12-07T09:06:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73\\\\n2025-12-07T09:06:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73 to /host/opt/cni/bin/\\\\n2025-12-07T09:06:55Z [verbose] multus-daemon started\\\\n2025-12-07T09:06:55Z [verbose] Readiness Indicator file check\\\\n2025-12-07T09:07:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.306389 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cba62b5c-dd17-4a93-922d-237e5a7557bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://286fc6c21ddb56cdcb4bdb001dbbbcbd294b23971933e39148c05efed583e5b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.324703 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.335778 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.335864 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.335884 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.335904 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.335947 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.340274 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.353635 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.375004 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.386709 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:42Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.439287 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.439358 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.439375 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.439395 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.439409 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.542013 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.542086 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.542130 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.542157 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.542175 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.613463 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:42 crc kubenswrapper[4838]: E1207 09:07:42.613584 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.644685 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.644720 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.644734 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.644753 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.644767 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.747242 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.747309 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.747332 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.747361 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.747383 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.849172 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.849227 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.849242 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.849262 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.849278 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.951480 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.951518 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.951528 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.951545 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:42 crc kubenswrapper[4838]: I1207 09:07:42.951559 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:42Z","lastTransitionTime":"2025-12-07T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.054030 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.054131 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.054173 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.054211 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.054236 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.157260 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.157372 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.157392 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.157421 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.157439 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.260135 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.260194 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.260211 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.260233 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.260250 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.362795 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.362895 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.362911 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.362935 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.362953 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.465882 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.466095 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.466137 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.466168 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.466198 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.569883 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.569980 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.570084 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.570110 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.570143 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.614381 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.614443 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.614463 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:43 crc kubenswrapper[4838]: E1207 09:07:43.614620 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:43 crc kubenswrapper[4838]: E1207 09:07:43.614758 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:43 crc kubenswrapper[4838]: E1207 09:07:43.614939 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.637963 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.659648 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.674465 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.674613 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.674670 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.674752 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.674852 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.677161 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.694198 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cba62b5c-dd17-4a93-922d-237e5a7557bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://286fc6c21ddb56cdcb4bdb001dbbbcbd294b23971933e39148c05efed583e5b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.714505 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.728370 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.742880 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.761351 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.777443 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.777931 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.778027 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.778094 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.778156 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.778212 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.799072 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:40Z\\\",\\\"message\\\":\\\"2025-12-07T09:06:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73\\\\n2025-12-07T09:06:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73 to /host/opt/cni/bin/\\\\n2025-12-07T09:06:55Z [verbose] multus-daemon started\\\\n2025-12-07T09:06:55Z [verbose] Readiness Indicator file check\\\\n2025-12-07T09:07:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.827672 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.841572 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.858144 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.874530 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.880490 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.880528 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.880539 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.880556 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.880568 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.891056 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.906121 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.923497 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:43Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.982551 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.982658 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.982724 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.982797 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:43 crc kubenswrapper[4838]: I1207 09:07:43.982902 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:43Z","lastTransitionTime":"2025-12-07T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.085635 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.085722 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.085741 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.085765 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.085782 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.188503 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.188554 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.188571 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.188593 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.188612 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.290908 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.290979 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.291002 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.291031 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.291053 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.393850 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.393908 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.393924 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.393946 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.393963 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.496272 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.496313 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.496324 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.496340 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.496352 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.598855 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.598934 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.598958 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.598994 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.599017 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.614040 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:44 crc kubenswrapper[4838]: E1207 09:07:44.614664 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.615277 4838 scope.go:117] "RemoveContainer" containerID="3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.702499 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.702572 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.702599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.702631 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.702654 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.717969 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.718200 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.718366 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.718505 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.718645 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: E1207 09:07:44.740309 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:44Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.746819 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.746913 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.746931 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.746951 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.746964 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: E1207 09:07:44.765409 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:44Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.770125 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.770159 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.770171 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.770186 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.770198 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: E1207 09:07:44.793133 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:44Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.797967 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.798054 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.798075 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.798130 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.798151 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: E1207 09:07:44.822329 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:44Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.828394 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.828438 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.828453 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.828474 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.828489 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: E1207 09:07:44.850110 4838 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98e1eaea-301b-4b2e-acd1-6718eb434b9d\\\",\\\"systemUUID\\\":\\\"db13a37f-60ef-4ad4-906a-e0d1f4e45870\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:44Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:44 crc kubenswrapper[4838]: E1207 09:07:44.850490 4838 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.852279 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.852300 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.852310 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.852325 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.852337 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.955339 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.955388 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.955396 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.955411 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:44 crc kubenswrapper[4838]: I1207 09:07:44.955421 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:44Z","lastTransitionTime":"2025-12-07T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.058228 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.058276 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.058288 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.058306 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.058318 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.097003 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/2.log" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.099944 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.100508 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.119395 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.131387 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.143620 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.152645 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.160549 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.160573 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.160583 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.160595 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.160603 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.163028 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.175619 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.188822 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.198646 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.206858 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.215965 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:40Z\\\",\\\"message\\\":\\\"2025-12-07T09:06:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73\\\\n2025-12-07T09:06:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73 to /host/opt/cni/bin/\\\\n2025-12-07T09:06:55Z [verbose] multus-daemon started\\\\n2025-12-07T09:06:55Z [verbose] Readiness Indicator file check\\\\n2025-12-07T09:07:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.225446 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cba62b5c-dd17-4a93-922d-237e5a7557bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://286fc6c21ddb56cdcb4bdb001dbbbcbd294b23971933e39148c05efed583e5b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.239401 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.252114 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.262621 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.262654 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.262663 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.262677 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.262686 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.263919 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.274100 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.291549 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.300302 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.364498 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.364532 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.364544 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.364560 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.364572 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.467879 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.467943 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.467966 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.467995 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.468016 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.571326 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.571401 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.571459 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.571489 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.571511 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.614512 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.614626 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:45 crc kubenswrapper[4838]: E1207 09:07:45.614694 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:45 crc kubenswrapper[4838]: E1207 09:07:45.614883 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.619561 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:45 crc kubenswrapper[4838]: E1207 09:07:45.619696 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.673884 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.673931 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.673944 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.673963 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.673976 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.776955 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.777001 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.777018 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.777040 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.777057 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.879347 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.879389 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.879398 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.879412 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.879424 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.981809 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.981886 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.981903 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.981932 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:45 crc kubenswrapper[4838]: I1207 09:07:45.981959 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:45Z","lastTransitionTime":"2025-12-07T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.083863 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.083926 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.083942 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.083969 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.083985 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.104479 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/3.log" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.105174 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/2.log" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.107645 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" exitCode=1 Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.107683 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.107720 4838 scope.go:117] "RemoveContainer" containerID="3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.108532 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:07:46 crc kubenswrapper[4838]: E1207 09:07:46.108695 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.125859 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.137769 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.150982 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.162983 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.174996 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.186741 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.186851 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.186877 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.186908 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.186931 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.189854 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.203364 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.225982 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.244875 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.264084 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.280879 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.290133 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.290180 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.290194 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.290212 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.290225 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.292910 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.306590 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:40Z\\\",\\\"message\\\":\\\"2025-12-07T09:06:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73\\\\n2025-12-07T09:06:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73 to /host/opt/cni/bin/\\\\n2025-12-07T09:06:55Z [verbose] multus-daemon started\\\\n2025-12-07T09:06:55Z [verbose] Readiness Indicator file check\\\\n2025-12-07T09:07:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.315963 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cba62b5c-dd17-4a93-922d-237e5a7557bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://286fc6c21ddb56cdcb4bdb001dbbbcbd294b23971933e39148c05efed583e5b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.327674 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.343545 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca78c1a13c33e5fcba99080e5c1b604d044ce15e3fe480cc43768ed74072ec7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:19Z\\\",\\\"message\\\":\\\"sFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.174\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:19.530605 6405 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-b26td\\\\nF1207 09:07:19.531849 6405 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:19Z is after 2025-08-24T17:21:41Z]\\\\nI1207 09:07:19.531456 6405 model_client.go:382] Update operations generated as: [{Op:update Ta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:45Z\\\",\\\"message\\\":\\\"ervices.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:45.506394 6774 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1207 09:07:45.506919 6774 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.352964 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:46Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.392615 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.392670 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.392685 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.392706 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.392723 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.495655 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.495710 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.495728 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.495752 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.495769 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.598751 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.598868 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.598894 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.598926 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.598952 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.614164 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:46 crc kubenswrapper[4838]: E1207 09:07:46.614354 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.701798 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.701868 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.701878 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.701894 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.701905 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.804636 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.804669 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.804678 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.804691 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.804702 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.906977 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.907011 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.907019 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.907031 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:46 crc kubenswrapper[4838]: I1207 09:07:46.907040 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:46Z","lastTransitionTime":"2025-12-07T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.009977 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.010078 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.010104 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.010137 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.010163 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.112765 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.112854 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.112873 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.112898 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.112917 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.114160 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/3.log" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.118687 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:07:47 crc kubenswrapper[4838]: E1207 09:07:47.118957 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.134654 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c033024a-e640-42b3-a9c2-534d367f0da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37a365d99edd8f059cf19862984e118cfdaa8c507137184e3ffb21134d5fe5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3598e5fb382931e946180e8b1ee0b7c02fcd96f4808e03f4c7b934dab8e086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ef26eb6359b70384a6e2de7346d855c07f4497393e892dc3009c11bddbca5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35aac13ba1a469e8df8d59026cc17cc87324dbdd10c850ca297fdbc2c904c01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.154229 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be24b21da5e7a4550f184603c8a772e0e87fdfb1d791c6101153e8bb6d599181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2148ee6ddf3c77f68db0c3d7a358d1ea9a60feb39a7b1fce5da3fa8e4889eabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.174755 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.186083 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rf7dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bf1dd6-0f35-4ec3-9d94-87800241f5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc607b63f4f0a9a0c0d8ad44876c23085bfd6b770fcd5ffa45c171debde03680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5xx2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rf7dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.201565 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1304517-3ac2-4aa3-afe2-01e27f87d880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ef9649f2b4c861f99b441ec149b0d62ef86728f499d6de4071f91609b4aac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tphnq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7v9qs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.216105 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.216141 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.216155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.216172 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.216186 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.216466 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7906ee0c9ec983465b5bdc2d3338ee21c209ea30a1b1cf66ea500cf93cf36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.232591 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5eb803b5-3546-42b8-829e-ba158fb6a2d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5855bd006fa8b03cac585e60a6307ba4853642c85ed24806d64e212f355f0cb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c15f4a68ed2a92290f022e6c2a4c3bcc32d188e982a478511befca992c02bc30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa24048af1c562e386c5cf67ecd53a78718d17456b875212dd52180bf155ed9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://181b346070652c7231a825fa924534d78f2b13d0f192dd48d9de10eb916d0610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48e7addec682506fbf60d5f540dca81d2a9d9f2f52a9421478027cd24c3f11e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5211d75546b4b6377328402d61088896663b7197758f741e60b76bac1ae36a2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec1cd6b1ed79f837a16614a43975acf1d813deeb4b05b9660be31dd8bf205dd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwpxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jkrnf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.244726 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9101534a-8a97-4d83-a97f-b8a0df78438d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca6e5cb761c03080991f20ad61d0d3bffefc3e4d8740a907c39544d238519fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ce063fd756e5115e3f342929f702662d94493502e814310653d7e92e8f9001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gljhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-b77qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.254008 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b26td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2ccf6c-bde8-4835-b4f5-5eaee2916835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23a3de33ef341a94020d7a4d3ac45cbaf820901dbb69d6811d90738b9eb76e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kt2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b26td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.266103 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bgkt9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:40Z\\\",\\\"message\\\":\\\"2025-12-07T09:06:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73\\\\n2025-12-07T09:06:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61d6a497-58dc-4e07-8a27-85ee8bf1cc73 to /host/opt/cni/bin/\\\\n2025-12-07T09:06:55Z [verbose] multus-daemon started\\\\n2025-12-07T09:06:55Z [verbose] Readiness Indicator file check\\\\n2025-12-07T09:07:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g6ln5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bgkt9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.274443 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cba62b5c-dd17-4a93-922d-237e5a7557bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://286fc6c21ddb56cdcb4bdb001dbbbcbd294b23971933e39148c05efed583e5b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d3fb10235a295dbc07cebac065b524798416891c2986589c797cbbaa0719a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.288169 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2c8a585-7f49-4169-81b8-34cb0d1a0be9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T09:06:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1207 09:06:37.037341 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 09:06:37.038516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3003158666/tls.crt::/tmp/serving-cert-3003158666/tls.key\\\\\\\"\\\\nI1207 09:06:46.655698 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 09:06:46.658266 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 09:06:46.658286 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 09:06:46.658310 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 09:06:46.658316 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 09:06:46.671677 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1207 09:06:46.671710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671714 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 09:06:46.671719 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 09:06:46.671725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 09:06:46.671728 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 09:06:46.671731 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1207 09:06:46.672043 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1207 09:06:46.675956 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.298471 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.308407 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.317331 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e428d9f078b9dfbd26f3388b7a2312d20ff02b98e6c84d6b90a1436ad398b931\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.320071 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.320114 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.320129 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.320147 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.320161 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.333622 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66f9850-4704-46f3-9745-bdf80caf8a98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T09:07:45Z\\\",\\\"message\\\":\\\"ervices.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1207 09:07:45.506394 6774 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1207 09:07:45.506919 6774 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T09:07:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T09:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T09:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T09:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcrtj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:06:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-brr6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.346486 4838 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T09:07:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c75kz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T09:07:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7nmsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T09:07:47Z is after 2025-08-24T17:21:41Z" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.423548 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.423611 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.423633 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.423660 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.423680 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.526058 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.526111 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.526126 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.526147 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.526162 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.614332 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.614374 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.614425 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:47 crc kubenswrapper[4838]: E1207 09:07:47.614549 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:47 crc kubenswrapper[4838]: E1207 09:07:47.614711 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:47 crc kubenswrapper[4838]: E1207 09:07:47.614912 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.629238 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.629295 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.629312 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.629333 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.629350 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.732475 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.732552 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.732576 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.732605 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.732626 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.835587 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.835647 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.835669 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.835700 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.835721 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.938641 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.938730 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.938751 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.938886 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:47 crc kubenswrapper[4838]: I1207 09:07:47.938934 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:47Z","lastTransitionTime":"2025-12-07T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.041306 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.041351 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.041363 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.041380 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.041392 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.144380 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.144870 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.144899 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.144931 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.144949 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.248190 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.248239 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.248256 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.248286 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.248303 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.350660 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.350711 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.350723 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.350738 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.350752 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.453233 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.453275 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.453286 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.453301 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.453312 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.555959 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.556033 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.556058 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.556090 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.556111 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.614067 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:48 crc kubenswrapper[4838]: E1207 09:07:48.614262 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.659359 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.659431 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.659455 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.659486 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.659506 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.762500 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.762574 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.762596 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.762627 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.762650 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.866004 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.866055 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.866074 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.866094 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.866110 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.968498 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.968917 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.969155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.969357 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:48 crc kubenswrapper[4838]: I1207 09:07:48.969537 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:48Z","lastTransitionTime":"2025-12-07T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.072004 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.072324 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.072468 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.072598 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.072723 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.175887 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.176235 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.176400 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.176545 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.176667 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.278996 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.279049 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.279064 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.279085 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.279099 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.382421 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.382479 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.382497 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.382524 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.382542 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.486053 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.486130 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.486150 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.486175 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.486193 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.589494 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.589530 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.589541 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.589554 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.589565 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.614189 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.614247 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.614199 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:49 crc kubenswrapper[4838]: E1207 09:07:49.614350 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:49 crc kubenswrapper[4838]: E1207 09:07:49.614444 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:49 crc kubenswrapper[4838]: E1207 09:07:49.614500 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.692690 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.692792 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.692809 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.692889 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.692916 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.796052 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.796122 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.796138 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.796164 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.796181 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.899767 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.899909 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.899931 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.899959 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:49 crc kubenswrapper[4838]: I1207 09:07:49.899980 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:49Z","lastTransitionTime":"2025-12-07T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.002133 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.002186 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.002204 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.002226 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.002243 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.104174 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.104209 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.104221 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.104235 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.104244 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.206907 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.206944 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.206955 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.206969 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.206980 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.309800 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.309861 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.309872 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.309889 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.309900 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.412397 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.412446 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.412458 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.412478 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.412492 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.514332 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.514385 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.514398 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.514417 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.514431 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.613452 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:50 crc kubenswrapper[4838]: E1207 09:07:50.613573 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.617021 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.617050 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.617058 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.617070 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.617107 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.720175 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.720286 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.720304 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.720342 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.720362 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.822769 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.822835 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.822845 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.822862 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.822875 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.926335 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.926408 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.926426 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.926452 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:50 crc kubenswrapper[4838]: I1207 09:07:50.926471 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:50Z","lastTransitionTime":"2025-12-07T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.029097 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.029150 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.029160 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.029177 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.029190 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.131327 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.131372 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.131409 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.131426 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.131438 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.233519 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.233567 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.233582 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.233601 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.233615 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.336978 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.337058 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.337086 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.337119 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.337141 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.434628 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.435036 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.435000381 +0000 UTC m=+152.142319438 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.439644 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.440065 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.440109 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.440151 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.440193 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.543548 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.543616 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.543808 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.543847 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.543860 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.614127 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.614243 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.615030 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.615130 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.615370 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.620180 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.636415 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.636472 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.636514 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.636539 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636623 4838 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636691 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.636667369 +0000 UTC m=+152.343986386 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636706 4838 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636715 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636765 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636787 4838 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636792 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.636763792 +0000 UTC m=+152.344082989 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636736 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636900 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.636866355 +0000 UTC m=+152.344185552 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636932 4838 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.636980 4838 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:51 crc kubenswrapper[4838]: E1207 09:07:51.637109 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.637079781 +0000 UTC m=+152.344398838 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.647610 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.647666 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.647686 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.647746 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.647766 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.750054 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.750118 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.750140 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.750168 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.750191 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.853599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.853629 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.853637 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.853649 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.853658 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.955471 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.955509 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.955518 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.955530 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:51 crc kubenswrapper[4838]: I1207 09:07:51.955539 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:51Z","lastTransitionTime":"2025-12-07T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.058306 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.058356 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.058433 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.058502 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.058520 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.160773 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.160832 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.160855 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.160876 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.160886 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.264270 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.264324 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.264341 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.264364 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.264380 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.367233 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.367271 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.367280 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.367297 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.367306 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.470304 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.470654 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.470865 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.471026 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.471343 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.574541 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.575101 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.575307 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.575492 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.575711 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.613781 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:52 crc kubenswrapper[4838]: E1207 09:07:52.614374 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.679190 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.679594 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.679759 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.679925 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.680059 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.783503 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.783559 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.783577 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.783598 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.783615 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.887111 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.887476 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.887640 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.887875 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.888089 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.991471 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.992131 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.992372 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.992601 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:52 crc kubenswrapper[4838]: I1207 09:07:52.992799 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:52Z","lastTransitionTime":"2025-12-07T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.095592 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.095631 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.095646 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.095664 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.095678 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.198202 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.198570 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.198784 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.199052 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.199266 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.302096 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.302166 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.302184 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.302207 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.302224 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.405173 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.405195 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.405203 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.405217 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.405226 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.507422 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.507470 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.507487 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.507508 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.507525 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.610684 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.610727 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.610738 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.610755 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.610769 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.613972 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:53 crc kubenswrapper[4838]: E1207 09:07:53.615885 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.616163 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.616219 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:53 crc kubenswrapper[4838]: E1207 09:07:53.616316 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:53 crc kubenswrapper[4838]: E1207 09:07:53.616807 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.639211 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=19.639182815 podStartE2EDuration="19.639182815s" podCreationTimestamp="2025-12-07 09:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.638905047 +0000 UTC m=+90.346224104" watchObservedRunningTime="2025-12-07 09:07:53.639182815 +0000 UTC m=+90.346501872" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.662472 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=66.662279845 podStartE2EDuration="1m6.662279845s" podCreationTimestamp="2025-12-07 09:06:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.66210223 +0000 UTC m=+90.369421307" watchObservedRunningTime="2025-12-07 09:07:53.662279845 +0000 UTC m=+90.369598882" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.716013 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.716217 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.717086 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.717618 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.717914 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.743882 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-b26td" podStartSLOduration=60.743860936 podStartE2EDuration="1m0.743860936s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.741299135 +0000 UTC m=+90.448618192" watchObservedRunningTime="2025-12-07 09:07:53.743860936 +0000 UTC m=+90.451179993" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.761584 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-bgkt9" podStartSLOduration=60.761569947 podStartE2EDuration="1m0.761569947s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.761196256 +0000 UTC m=+90.468515273" watchObservedRunningTime="2025-12-07 09:07:53.761569947 +0000 UTC m=+90.468888964" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.821340 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.821369 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.821379 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.821408 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.821417 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.857104 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=38.857085574 podStartE2EDuration="38.857085574s" podCreationTimestamp="2025-12-07 09:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.840580126 +0000 UTC m=+90.547899153" watchObservedRunningTime="2025-12-07 09:07:53.857085574 +0000 UTC m=+90.564404591" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.878081 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rf7dv" podStartSLOduration=60.878061775 podStartE2EDuration="1m0.878061775s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.877674374 +0000 UTC m=+90.584993391" watchObservedRunningTime="2025-12-07 09:07:53.878061775 +0000 UTC m=+90.585380792" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.888463 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podStartSLOduration=60.888448483 podStartE2EDuration="1m0.888448483s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.888384131 +0000 UTC m=+90.595703148" watchObservedRunningTime="2025-12-07 09:07:53.888448483 +0000 UTC m=+90.595767500" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.923289 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.923320 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.923328 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.923341 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.923350 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:53Z","lastTransitionTime":"2025-12-07T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.929080 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jkrnf" podStartSLOduration=60.929064538 podStartE2EDuration="1m0.929064538s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.928482392 +0000 UTC m=+90.635801409" watchObservedRunningTime="2025-12-07 09:07:53.929064538 +0000 UTC m=+90.636383555" Dec 07 09:07:53 crc kubenswrapper[4838]: I1207 09:07:53.941244 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-b77qb" podStartSLOduration=59.941229196 podStartE2EDuration="59.941229196s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:53.940747732 +0000 UTC m=+90.648066749" watchObservedRunningTime="2025-12-07 09:07:53.941229196 +0000 UTC m=+90.648548213" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.025392 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.025439 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.025452 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.025469 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.025483 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.128113 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.128155 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.128164 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.128178 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.128188 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.230628 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.230673 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.230686 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.230702 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.230714 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.334168 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.336921 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.336961 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.336989 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.337011 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.440357 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.440944 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.440966 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.440979 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.440989 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.544283 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.544332 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.544349 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.544371 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.544386 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.614232 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:54 crc kubenswrapper[4838]: E1207 09:07:54.614731 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.647620 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.647687 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.647710 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.647740 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.647765 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.750302 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.750341 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.750352 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.750368 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.750381 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.852268 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.852353 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.852370 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.852390 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.852405 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.955599 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.955673 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.955698 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.955730 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:54 crc kubenswrapper[4838]: I1207 09:07:54.955754 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:54Z","lastTransitionTime":"2025-12-07T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.058710 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.059061 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.059148 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.059261 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.059342 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:55Z","lastTransitionTime":"2025-12-07T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.162545 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.162790 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.162901 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.162991 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.163118 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:55Z","lastTransitionTime":"2025-12-07T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.213574 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.213629 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.213648 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.213671 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.213690 4838 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T09:07:55Z","lastTransitionTime":"2025-12-07T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.280723 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8"] Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.281342 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.283693 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.283775 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.285580 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.285796 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.379673 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d97a85-bd97-4add-915e-a55d3a1c2b28-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.379804 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/41d97a85-bd97-4add-915e-a55d3a1c2b28-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.379961 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41d97a85-bd97-4add-915e-a55d3a1c2b28-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.380018 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/41d97a85-bd97-4add-915e-a55d3a1c2b28-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.380116 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41d97a85-bd97-4add-915e-a55d3a1c2b28-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.481450 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/41d97a85-bd97-4add-915e-a55d3a1c2b28-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.481589 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41d97a85-bd97-4add-915e-a55d3a1c2b28-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.481599 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/41d97a85-bd97-4add-915e-a55d3a1c2b28-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.481632 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d97a85-bd97-4add-915e-a55d3a1c2b28-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.481717 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/41d97a85-bd97-4add-915e-a55d3a1c2b28-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.481767 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41d97a85-bd97-4add-915e-a55d3a1c2b28-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.482170 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/41d97a85-bd97-4add-915e-a55d3a1c2b28-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.483522 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d97a85-bd97-4add-915e-a55d3a1c2b28-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.492010 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41d97a85-bd97-4add-915e-a55d3a1c2b28-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.512994 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41d97a85-bd97-4add-915e-a55d3a1c2b28-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xrlx8\" (UID: \"41d97a85-bd97-4add-915e-a55d3a1c2b28\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.594553 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.614365 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.614385 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:55 crc kubenswrapper[4838]: I1207 09:07:55.614449 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:55 crc kubenswrapper[4838]: E1207 09:07:55.615221 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:55 crc kubenswrapper[4838]: E1207 09:07:55.615319 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:55 crc kubenswrapper[4838]: E1207 09:07:55.615484 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:55 crc kubenswrapper[4838]: W1207 09:07:55.619888 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41d97a85_bd97_4add_915e_a55d3a1c2b28.slice/crio-cf04d58c9c79d55d41918eb7af7d4d0412962f930a04686826f1fe793688d16f WatchSource:0}: Error finding container cf04d58c9c79d55d41918eb7af7d4d0412962f930a04686826f1fe793688d16f: Status 404 returned error can't find the container with id cf04d58c9c79d55d41918eb7af7d4d0412962f930a04686826f1fe793688d16f Dec 07 09:07:56 crc kubenswrapper[4838]: I1207 09:07:56.154920 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" event={"ID":"41d97a85-bd97-4add-915e-a55d3a1c2b28","Type":"ContainerStarted","Data":"cf04d58c9c79d55d41918eb7af7d4d0412962f930a04686826f1fe793688d16f"} Dec 07 09:07:56 crc kubenswrapper[4838]: I1207 09:07:56.613593 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:56 crc kubenswrapper[4838]: E1207 09:07:56.613768 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:57 crc kubenswrapper[4838]: I1207 09:07:57.158946 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" event={"ID":"41d97a85-bd97-4add-915e-a55d3a1c2b28","Type":"ContainerStarted","Data":"aeb7c4be8b1605e1f41c71a2a5b8e8f24bfd87b184b45861f5368947196d92d5"} Dec 07 09:07:57 crc kubenswrapper[4838]: I1207 09:07:57.177193 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xrlx8" podStartSLOduration=64.177173172 podStartE2EDuration="1m4.177173172s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:07:57.175488255 +0000 UTC m=+93.882807312" watchObservedRunningTime="2025-12-07 09:07:57.177173172 +0000 UTC m=+93.884492199" Dec 07 09:07:57 crc kubenswrapper[4838]: I1207 09:07:57.615129 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:57 crc kubenswrapper[4838]: I1207 09:07:57.615598 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:57 crc kubenswrapper[4838]: E1207 09:07:57.615760 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:57 crc kubenswrapper[4838]: I1207 09:07:57.616175 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:57 crc kubenswrapper[4838]: E1207 09:07:57.616279 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:07:57 crc kubenswrapper[4838]: E1207 09:07:57.616538 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:57 crc kubenswrapper[4838]: I1207 09:07:57.630472 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 07 09:07:58 crc kubenswrapper[4838]: I1207 09:07:58.613672 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:07:58 crc kubenswrapper[4838]: E1207 09:07:58.614273 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:07:59 crc kubenswrapper[4838]: I1207 09:07:59.614573 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:07:59 crc kubenswrapper[4838]: I1207 09:07:59.614606 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:07:59 crc kubenswrapper[4838]: I1207 09:07:59.614569 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:07:59 crc kubenswrapper[4838]: E1207 09:07:59.614711 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:07:59 crc kubenswrapper[4838]: E1207 09:07:59.614837 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:07:59 crc kubenswrapper[4838]: E1207 09:07:59.614908 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:00 crc kubenswrapper[4838]: I1207 09:08:00.614342 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:00 crc kubenswrapper[4838]: E1207 09:08:00.614742 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:01 crc kubenswrapper[4838]: I1207 09:08:01.614520 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:01 crc kubenswrapper[4838]: I1207 09:08:01.614553 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:01 crc kubenswrapper[4838]: E1207 09:08:01.614740 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:01 crc kubenswrapper[4838]: E1207 09:08:01.615014 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:01 crc kubenswrapper[4838]: I1207 09:08:01.615672 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:01 crc kubenswrapper[4838]: E1207 09:08:01.615783 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:01 crc kubenswrapper[4838]: I1207 09:08:01.633858 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 07 09:08:02 crc kubenswrapper[4838]: I1207 09:08:02.614477 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:02 crc kubenswrapper[4838]: E1207 09:08:02.615486 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:02 crc kubenswrapper[4838]: I1207 09:08:02.616130 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:08:02 crc kubenswrapper[4838]: E1207 09:08:02.616692 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:08:03 crc kubenswrapper[4838]: I1207 09:08:03.614309 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:03 crc kubenswrapper[4838]: I1207 09:08:03.614423 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:03 crc kubenswrapper[4838]: E1207 09:08:03.619746 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:03 crc kubenswrapper[4838]: I1207 09:08:03.619785 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:03 crc kubenswrapper[4838]: E1207 09:08:03.623167 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:03 crc kubenswrapper[4838]: E1207 09:08:03.623453 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:03 crc kubenswrapper[4838]: I1207 09:08:03.679331 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.679307284 podStartE2EDuration="6.679307284s" podCreationTimestamp="2025-12-07 09:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:03.677556025 +0000 UTC m=+100.384875112" watchObservedRunningTime="2025-12-07 09:08:03.679307284 +0000 UTC m=+100.386626331" Dec 07 09:08:04 crc kubenswrapper[4838]: I1207 09:08:04.613852 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:04 crc kubenswrapper[4838]: E1207 09:08:04.614070 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:05 crc kubenswrapper[4838]: I1207 09:08:05.614561 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:05 crc kubenswrapper[4838]: I1207 09:08:05.614649 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:05 crc kubenswrapper[4838]: I1207 09:08:05.614604 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:05 crc kubenswrapper[4838]: E1207 09:08:05.614781 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:05 crc kubenswrapper[4838]: E1207 09:08:05.614996 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:05 crc kubenswrapper[4838]: E1207 09:08:05.615176 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:06 crc kubenswrapper[4838]: I1207 09:08:06.614246 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:06 crc kubenswrapper[4838]: E1207 09:08:06.614407 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:07 crc kubenswrapper[4838]: I1207 09:08:07.614378 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:07 crc kubenswrapper[4838]: I1207 09:08:07.614477 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:07 crc kubenswrapper[4838]: E1207 09:08:07.614622 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:07 crc kubenswrapper[4838]: E1207 09:08:07.614691 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:07 crc kubenswrapper[4838]: I1207 09:08:07.615693 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:07 crc kubenswrapper[4838]: E1207 09:08:07.616077 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:08 crc kubenswrapper[4838]: I1207 09:08:08.614323 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:08 crc kubenswrapper[4838]: E1207 09:08:08.614783 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:09 crc kubenswrapper[4838]: I1207 09:08:09.613532 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:09 crc kubenswrapper[4838]: I1207 09:08:09.613586 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:09 crc kubenswrapper[4838]: E1207 09:08:09.613633 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:09 crc kubenswrapper[4838]: E1207 09:08:09.613697 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:09 crc kubenswrapper[4838]: I1207 09:08:09.613536 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:09 crc kubenswrapper[4838]: E1207 09:08:09.613808 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:10 crc kubenswrapper[4838]: I1207 09:08:10.613494 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:10 crc kubenswrapper[4838]: E1207 09:08:10.613666 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:11 crc kubenswrapper[4838]: I1207 09:08:11.613516 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:11 crc kubenswrapper[4838]: I1207 09:08:11.613611 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:11 crc kubenswrapper[4838]: I1207 09:08:11.613517 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:11 crc kubenswrapper[4838]: E1207 09:08:11.613685 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:11 crc kubenswrapper[4838]: E1207 09:08:11.613744 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:11 crc kubenswrapper[4838]: E1207 09:08:11.613873 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:12 crc kubenswrapper[4838]: I1207 09:08:12.339514 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:12 crc kubenswrapper[4838]: E1207 09:08:12.339736 4838 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:08:12 crc kubenswrapper[4838]: E1207 09:08:12.339847 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs podName:36f02276-0fd8-4d6e-9b23-9aeb7068d8ff nodeName:}" failed. No retries permitted until 2025-12-07 09:09:16.339788249 +0000 UTC m=+173.047107306 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs") pod "network-metrics-daemon-7nmsx" (UID: "36f02276-0fd8-4d6e-9b23-9aeb7068d8ff") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 09:08:12 crc kubenswrapper[4838]: I1207 09:08:12.613946 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:12 crc kubenswrapper[4838]: E1207 09:08:12.614141 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:13 crc kubenswrapper[4838]: I1207 09:08:13.614429 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:13 crc kubenswrapper[4838]: I1207 09:08:13.614597 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:13 crc kubenswrapper[4838]: E1207 09:08:13.616545 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:13 crc kubenswrapper[4838]: E1207 09:08:13.616721 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:13 crc kubenswrapper[4838]: I1207 09:08:13.617197 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:13 crc kubenswrapper[4838]: E1207 09:08:13.617418 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:14 crc kubenswrapper[4838]: I1207 09:08:14.613944 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:14 crc kubenswrapper[4838]: E1207 09:08:14.614163 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:15 crc kubenswrapper[4838]: I1207 09:08:15.614984 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:15 crc kubenswrapper[4838]: I1207 09:08:15.615033 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:15 crc kubenswrapper[4838]: E1207 09:08:15.615104 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:15 crc kubenswrapper[4838]: I1207 09:08:15.615208 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:15 crc kubenswrapper[4838]: E1207 09:08:15.615348 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:15 crc kubenswrapper[4838]: E1207 09:08:15.615438 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:16 crc kubenswrapper[4838]: I1207 09:08:16.614591 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:16 crc kubenswrapper[4838]: E1207 09:08:16.614738 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:17 crc kubenswrapper[4838]: I1207 09:08:17.614088 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:17 crc kubenswrapper[4838]: I1207 09:08:17.614132 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:17 crc kubenswrapper[4838]: E1207 09:08:17.614246 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:17 crc kubenswrapper[4838]: I1207 09:08:17.614319 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:17 crc kubenswrapper[4838]: E1207 09:08:17.614472 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:17 crc kubenswrapper[4838]: E1207 09:08:17.614781 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:17 crc kubenswrapper[4838]: I1207 09:08:17.616034 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:08:17 crc kubenswrapper[4838]: E1207 09:08:17.616329 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-brr6g_openshift-ovn-kubernetes(a66f9850-4704-46f3-9745-bdf80caf8a98)\"" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" Dec 07 09:08:18 crc kubenswrapper[4838]: I1207 09:08:18.614325 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:18 crc kubenswrapper[4838]: E1207 09:08:18.614485 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:19 crc kubenswrapper[4838]: I1207 09:08:19.613601 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:19 crc kubenswrapper[4838]: I1207 09:08:19.613680 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:19 crc kubenswrapper[4838]: E1207 09:08:19.613754 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:19 crc kubenswrapper[4838]: I1207 09:08:19.613601 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:19 crc kubenswrapper[4838]: E1207 09:08:19.614012 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:19 crc kubenswrapper[4838]: E1207 09:08:19.614133 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:20 crc kubenswrapper[4838]: I1207 09:08:20.614378 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:20 crc kubenswrapper[4838]: E1207 09:08:20.614576 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:21 crc kubenswrapper[4838]: I1207 09:08:21.613729 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:21 crc kubenswrapper[4838]: E1207 09:08:21.613919 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:21 crc kubenswrapper[4838]: I1207 09:08:21.614174 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:21 crc kubenswrapper[4838]: E1207 09:08:21.614278 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:21 crc kubenswrapper[4838]: I1207 09:08:21.614388 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:21 crc kubenswrapper[4838]: E1207 09:08:21.614599 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:22 crc kubenswrapper[4838]: I1207 09:08:22.613650 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:22 crc kubenswrapper[4838]: E1207 09:08:22.613961 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:23 crc kubenswrapper[4838]: E1207 09:08:23.574921 4838 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 07 09:08:23 crc kubenswrapper[4838]: I1207 09:08:23.614473 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:23 crc kubenswrapper[4838]: I1207 09:08:23.618570 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:23 crc kubenswrapper[4838]: I1207 09:08:23.618616 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:23 crc kubenswrapper[4838]: E1207 09:08:23.618691 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:23 crc kubenswrapper[4838]: E1207 09:08:23.618754 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:23 crc kubenswrapper[4838]: E1207 09:08:23.618635 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:23 crc kubenswrapper[4838]: E1207 09:08:23.987909 4838 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 09:08:24 crc kubenswrapper[4838]: I1207 09:08:24.614234 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:24 crc kubenswrapper[4838]: E1207 09:08:24.614413 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:25 crc kubenswrapper[4838]: I1207 09:08:25.614200 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:25 crc kubenswrapper[4838]: I1207 09:08:25.614199 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:25 crc kubenswrapper[4838]: E1207 09:08:25.614354 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:25 crc kubenswrapper[4838]: E1207 09:08:25.614772 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:25 crc kubenswrapper[4838]: I1207 09:08:25.615317 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:25 crc kubenswrapper[4838]: E1207 09:08:25.615579 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:26 crc kubenswrapper[4838]: I1207 09:08:26.614326 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:26 crc kubenswrapper[4838]: E1207 09:08:26.614598 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.266996 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/1.log" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.267664 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/0.log" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.267733 4838 generic.go:334] "Generic (PLEG): container finished" podID="ca7ddad3-e3c1-47c4-b4a8-d332f9d69227" containerID="6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9" exitCode=1 Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.267778 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerDied","Data":"6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9"} Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.267883 4838 scope.go:117] "RemoveContainer" containerID="bbdb6973b6103590b9d9e1033207c5d45177a7784206aac332a1be34487ed84b" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.269407 4838 scope.go:117] "RemoveContainer" containerID="6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9" Dec 07 09:08:27 crc kubenswrapper[4838]: E1207 09:08:27.269799 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-bgkt9_openshift-multus(ca7ddad3-e3c1-47c4-b4a8-d332f9d69227)\"" pod="openshift-multus/multus-bgkt9" podUID="ca7ddad3-e3c1-47c4-b4a8-d332f9d69227" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.312925 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=26.312893261 podStartE2EDuration="26.312893261s" podCreationTimestamp="2025-12-07 09:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:03.702959569 +0000 UTC m=+100.410278626" watchObservedRunningTime="2025-12-07 09:08:27.312893261 +0000 UTC m=+124.020212338" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.614292 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:27 crc kubenswrapper[4838]: E1207 09:08:27.614505 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.614992 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:27 crc kubenswrapper[4838]: I1207 09:08:27.615073 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:27 crc kubenswrapper[4838]: E1207 09:08:27.615141 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:27 crc kubenswrapper[4838]: E1207 09:08:27.615244 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:28 crc kubenswrapper[4838]: I1207 09:08:28.274330 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/1.log" Dec 07 09:08:28 crc kubenswrapper[4838]: I1207 09:08:28.613955 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:28 crc kubenswrapper[4838]: E1207 09:08:28.614123 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:28 crc kubenswrapper[4838]: E1207 09:08:28.989059 4838 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 09:08:29 crc kubenswrapper[4838]: I1207 09:08:29.614291 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:29 crc kubenswrapper[4838]: I1207 09:08:29.614408 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:29 crc kubenswrapper[4838]: E1207 09:08:29.614449 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:29 crc kubenswrapper[4838]: I1207 09:08:29.614592 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:29 crc kubenswrapper[4838]: E1207 09:08:29.614808 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:29 crc kubenswrapper[4838]: E1207 09:08:29.615298 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:30 crc kubenswrapper[4838]: I1207 09:08:30.614126 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:30 crc kubenswrapper[4838]: E1207 09:08:30.614343 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:31 crc kubenswrapper[4838]: I1207 09:08:31.613959 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:31 crc kubenswrapper[4838]: I1207 09:08:31.614034 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:31 crc kubenswrapper[4838]: I1207 09:08:31.613963 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:31 crc kubenswrapper[4838]: E1207 09:08:31.614166 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:31 crc kubenswrapper[4838]: E1207 09:08:31.614362 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:31 crc kubenswrapper[4838]: E1207 09:08:31.614445 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:31 crc kubenswrapper[4838]: I1207 09:08:31.615003 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:08:32 crc kubenswrapper[4838]: I1207 09:08:32.291402 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/3.log" Dec 07 09:08:32 crc kubenswrapper[4838]: I1207 09:08:32.294615 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerStarted","Data":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} Dec 07 09:08:32 crc kubenswrapper[4838]: I1207 09:08:32.296457 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:08:32 crc kubenswrapper[4838]: I1207 09:08:32.330384 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podStartSLOduration=99.330368758 podStartE2EDuration="1m39.330368758s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:32.32862291 +0000 UTC m=+129.035941927" watchObservedRunningTime="2025-12-07 09:08:32.330368758 +0000 UTC m=+129.037687775" Dec 07 09:08:32 crc kubenswrapper[4838]: I1207 09:08:32.469057 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7nmsx"] Dec 07 09:08:32 crc kubenswrapper[4838]: I1207 09:08:32.469203 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:32 crc kubenswrapper[4838]: E1207 09:08:32.469331 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:32 crc kubenswrapper[4838]: I1207 09:08:32.613963 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:32 crc kubenswrapper[4838]: E1207 09:08:32.614075 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:33 crc kubenswrapper[4838]: I1207 09:08:33.614108 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:33 crc kubenswrapper[4838]: I1207 09:08:33.614319 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:33 crc kubenswrapper[4838]: E1207 09:08:33.615915 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:33 crc kubenswrapper[4838]: E1207 09:08:33.616335 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:33 crc kubenswrapper[4838]: E1207 09:08:33.989896 4838 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 09:08:34 crc kubenswrapper[4838]: I1207 09:08:34.614025 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:34 crc kubenswrapper[4838]: I1207 09:08:34.614052 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:34 crc kubenswrapper[4838]: E1207 09:08:34.614223 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:34 crc kubenswrapper[4838]: E1207 09:08:34.614290 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:35 crc kubenswrapper[4838]: I1207 09:08:35.613561 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:35 crc kubenswrapper[4838]: I1207 09:08:35.613668 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:35 crc kubenswrapper[4838]: E1207 09:08:35.613742 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:35 crc kubenswrapper[4838]: E1207 09:08:35.613962 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:36 crc kubenswrapper[4838]: I1207 09:08:36.614083 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:36 crc kubenswrapper[4838]: E1207 09:08:36.614237 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:36 crc kubenswrapper[4838]: I1207 09:08:36.614091 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:36 crc kubenswrapper[4838]: E1207 09:08:36.614676 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:37 crc kubenswrapper[4838]: I1207 09:08:37.614513 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:37 crc kubenswrapper[4838]: I1207 09:08:37.614528 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:37 crc kubenswrapper[4838]: E1207 09:08:37.614691 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:37 crc kubenswrapper[4838]: E1207 09:08:37.614857 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:38 crc kubenswrapper[4838]: I1207 09:08:38.614017 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:38 crc kubenswrapper[4838]: I1207 09:08:38.614060 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:38 crc kubenswrapper[4838]: E1207 09:08:38.614297 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:38 crc kubenswrapper[4838]: E1207 09:08:38.614434 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:38 crc kubenswrapper[4838]: E1207 09:08:38.991705 4838 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 09:08:39 crc kubenswrapper[4838]: I1207 09:08:39.613709 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:39 crc kubenswrapper[4838]: I1207 09:08:39.613710 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:39 crc kubenswrapper[4838]: E1207 09:08:39.614014 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:39 crc kubenswrapper[4838]: E1207 09:08:39.614213 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:40 crc kubenswrapper[4838]: I1207 09:08:40.613639 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:40 crc kubenswrapper[4838]: I1207 09:08:40.613678 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:40 crc kubenswrapper[4838]: E1207 09:08:40.613810 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:40 crc kubenswrapper[4838]: E1207 09:08:40.614040 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:41 crc kubenswrapper[4838]: I1207 09:08:41.614601 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:41 crc kubenswrapper[4838]: I1207 09:08:41.614941 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:41 crc kubenswrapper[4838]: E1207 09:08:41.615031 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:41 crc kubenswrapper[4838]: I1207 09:08:41.615142 4838 scope.go:117] "RemoveContainer" containerID="6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9" Dec 07 09:08:41 crc kubenswrapper[4838]: E1207 09:08:41.615275 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:42 crc kubenswrapper[4838]: I1207 09:08:42.328915 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/1.log" Dec 07 09:08:42 crc kubenswrapper[4838]: I1207 09:08:42.329211 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerStarted","Data":"5d011f28192f6e3f2d937b0835d48b047eb58e40d4a7df64368b680dcf332d8e"} Dec 07 09:08:42 crc kubenswrapper[4838]: I1207 09:08:42.614018 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:42 crc kubenswrapper[4838]: I1207 09:08:42.614094 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:42 crc kubenswrapper[4838]: E1207 09:08:42.614207 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 09:08:42 crc kubenswrapper[4838]: E1207 09:08:42.614503 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7nmsx" podUID="36f02276-0fd8-4d6e-9b23-9aeb7068d8ff" Dec 07 09:08:43 crc kubenswrapper[4838]: I1207 09:08:43.614388 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:43 crc kubenswrapper[4838]: E1207 09:08:43.617697 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 09:08:43 crc kubenswrapper[4838]: I1207 09:08:43.617777 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:43 crc kubenswrapper[4838]: E1207 09:08:43.618138 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 09:08:44 crc kubenswrapper[4838]: I1207 09:08:44.613790 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:44 crc kubenswrapper[4838]: I1207 09:08:44.613896 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:08:44 crc kubenswrapper[4838]: I1207 09:08:44.615713 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 07 09:08:44 crc kubenswrapper[4838]: I1207 09:08:44.616154 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 07 09:08:44 crc kubenswrapper[4838]: I1207 09:08:44.616559 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 07 09:08:44 crc kubenswrapper[4838]: I1207 09:08:44.616744 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 07 09:08:45 crc kubenswrapper[4838]: I1207 09:08:45.613619 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:45 crc kubenswrapper[4838]: I1207 09:08:45.613972 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:45 crc kubenswrapper[4838]: I1207 09:08:45.617352 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 07 09:08:45 crc kubenswrapper[4838]: I1207 09:08:45.622200 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.339849 4838 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.381931 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ng7p6"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.383032 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.384933 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.385351 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.387931 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wmz9f"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.388434 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.388986 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qll5w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.389682 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.390427 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.391050 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.391613 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.391909 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.393693 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.394113 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.394223 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.395600 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.395721 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.395873 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.401786 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.402147 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.404093 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.404316 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.406906 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.413214 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.413639 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.413761 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.414017 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.414084 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.414128 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.414209 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.414529 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.433840 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wmz9f"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.446147 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.447288 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.447509 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.448077 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.448310 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.448355 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.454497 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.454712 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.454859 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.454960 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455055 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455155 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455314 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455363 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455407 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455495 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455517 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455620 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455636 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455698 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455730 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455772 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455787 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.455329 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.456738 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.456939 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.457024 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.457694 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qll5w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.458118 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.459154 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4fqk"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.459897 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.463931 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.464248 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.464266 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.474171 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ng7p6"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.474229 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.474632 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.475045 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.478021 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.478395 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.480322 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-xrdpm"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.480832 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.481117 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.481170 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.481575 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.481676 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.481962 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.482220 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.482788 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.482984 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.483206 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.483417 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.483592 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.484175 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.487898 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dckpn"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.488296 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4fqk"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.488317 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-c8jw7"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.488723 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.489125 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.489623 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.489681 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4zsg8"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.491142 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.492229 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.494036 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.499340 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.502578 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-88r9w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.502919 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.503845 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.504065 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.504309 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.504690 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.504834 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.494220 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.513181 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.513878 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.514149 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.514484 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.515761 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.499046 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517050 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9038da7b-0b0e-44f6-abda-827d7de86d0d-audit-dir\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517084 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq68w\" (UniqueName: \"kubernetes.io/projected/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-kube-api-access-zq68w\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517109 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzhml\" (UniqueName: \"kubernetes.io/projected/e432ea72-cff7-4826-8001-125443ca158a-kube-api-access-mzhml\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517233 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950f31f5-e378-4a18-96f7-04e2091e2ca3-serving-cert\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517276 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-888xr\" (UniqueName: \"kubernetes.io/projected/950f31f5-e378-4a18-96f7-04e2091e2ca3-kube-api-access-888xr\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.499104 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517396 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-audit-dir\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.505007 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517560 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-config\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517584 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn74z\" (UniqueName: \"kubernetes.io/projected/15119df3-c69c-4ee2-8a7a-f9373494178d-kube-api-access-dn74z\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517739 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-client-ca\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517786 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517772 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-image-import-ca\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517884 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.517986 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-config\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518214 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15119df3-c69c-4ee2-8a7a-f9373494178d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518319 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-encryption-config\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518370 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518401 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-etcd-client\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518453 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-config\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518535 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9038da7b-0b0e-44f6-abda-827d7de86d0d-node-pullsecrets\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518604 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-serving-cert\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518628 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-etcd-serving-ca\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518686 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15119df3-c69c-4ee2-8a7a-f9373494178d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518794 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.518949 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e432ea72-cff7-4826-8001-125443ca158a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519028 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvxgm\" (UniqueName: \"kubernetes.io/projected/9b38416c-b5e5-4195-b950-6e449e035c8b-kube-api-access-cvxgm\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519051 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-etcd-client\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519113 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-audit-policies\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519164 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b38416c-b5e5-4195-b950-6e449e035c8b-serving-cert\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519222 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519258 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519387 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-encryption-config\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519434 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e432ea72-cff7-4826-8001-125443ca158a-images\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519509 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-serving-cert\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519540 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-client-ca\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519623 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-audit\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519770 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kd4z\" (UniqueName: \"kubernetes.io/projected/9038da7b-0b0e-44f6-abda-827d7de86d0d-kube-api-access-8kd4z\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.519832 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e432ea72-cff7-4826-8001-125443ca158a-config\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.520647 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.521118 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.532842 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.533157 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.533273 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.533828 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.534241 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.534522 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.534959 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535021 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535055 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535137 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535227 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535239 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535326 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535491 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535634 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535769 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.535880 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.536024 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.536145 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.536271 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.536424 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.536503 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.536165 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.537038 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.538195 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-pvp5p"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.538269 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.538536 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.538657 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.539494 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.542776 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.543772 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.545503 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.545761 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.548593 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.550985 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.553509 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.555434 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-7qwhs"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.555932 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.556402 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.556740 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gwtm7"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.557069 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-srd6x"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.557415 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xrdpm"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.557664 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-srd6x" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.557944 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.558054 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.558159 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.557998 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.558629 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.558754 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.561213 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-66ndx"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.561304 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.561780 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.570398 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.570709 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.578416 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.578604 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.594115 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.614229 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.614936 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-htg8d"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.615422 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.615890 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.618979 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.619023 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.619342 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.619357 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.620763 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-default-certificate\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.621826 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95bf0de9-2c4f-4504-a083-49cda7af2c1a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.621953 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ff7n\" (UniqueName: \"kubernetes.io/projected/df3e0f1c-d750-4b75-8f31-aa4faf37093c-kube-api-access-8ff7n\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622031 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-config\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622097 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/860ad6f7-00fc-4974-a703-389c3ca9d3a0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622170 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622246 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4de2ff5d-8c21-4412-b175-446ec4eec272-serving-cert\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622311 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6gnl\" (UniqueName: \"kubernetes.io/projected/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-kube-api-access-p6gnl\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622385 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvxgm\" (UniqueName: \"kubernetes.io/projected/9b38416c-b5e5-4195-b950-6e449e035c8b-kube-api-access-cvxgm\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622460 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-etcd-client\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622543 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-audit-policies\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622622 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b38416c-b5e5-4195-b950-6e449e035c8b-serving-cert\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622688 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-config\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622752 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1fbf61d9-c3ca-4298-aa49-8f113925aec7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s6wmc\" (UID: \"1fbf61d9-c3ca-4298-aa49-8f113925aec7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622861 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8cecca3b-390d-4346-9bb4-586c9b7f56ea-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.622942 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.623011 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-encryption-config\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.625477 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.625651 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-serving-cert\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.625737 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-client-ca\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.625822 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.625893 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-auth-proxy-config\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.625958 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-config\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.626025 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e432ea72-cff7-4826-8001-125443ca158a-images\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.626100 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-audit\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.627030 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kd4z\" (UniqueName: \"kubernetes.io/projected/9038da7b-0b0e-44f6-abda-827d7de86d0d-kube-api-access-8kd4z\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.627127 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e432ea72-cff7-4826-8001-125443ca158a-config\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.627619 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-client-ca\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.623558 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.628240 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll54w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.628866 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-etcd-client\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.629724 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e432ea72-cff7-4826-8001-125443ca158a-images\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.629790 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-plzvk"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.630257 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.630606 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-audit\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.623620 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.630727 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.630743 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.627036 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.631014 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.625078 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.631185 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.631361 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.632852 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.632971 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e432ea72-cff7-4826-8001-125443ca158a-config\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.633157 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.633883 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09c2b041-5634-431c-bc61-5a4418c5296e-audit-dir\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.633911 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95bf0de9-2c4f-4504-a083-49cda7af2c1a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.633933 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nxc9\" (UniqueName: \"kubernetes.io/projected/eb898d9c-aa97-4aa4-a31f-ab8659120558-kube-api-access-2nxc9\") pod \"dns-operator-744455d44c-c8jw7\" (UID: \"eb898d9c-aa97-4aa4-a31f-ab8659120558\") " pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.633954 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9038da7b-0b0e-44f6-abda-827d7de86d0d-audit-dir\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.633969 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq68w\" (UniqueName: \"kubernetes.io/projected/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-kube-api-access-zq68w\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.633987 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-serving-cert\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634003 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-service-ca\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634035 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzhml\" (UniqueName: \"kubernetes.io/projected/e432ea72-cff7-4826-8001-125443ca158a-kube-api-access-mzhml\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634049 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eb898d9c-aa97-4aa4-a31f-ab8659120558-metrics-tls\") pod \"dns-operator-744455d44c-c8jw7\" (UID: \"eb898d9c-aa97-4aa4-a31f-ab8659120558\") " pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634067 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwt82\" (UniqueName: \"kubernetes.io/projected/1fbf61d9-c3ca-4298-aa49-8f113925aec7-kube-api-access-nwt82\") pod \"cluster-samples-operator-665b6dd947-s6wmc\" (UID: \"1fbf61d9-c3ca-4298-aa49-8f113925aec7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634083 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/860ad6f7-00fc-4974-a703-389c3ca9d3a0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634098 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66r7p\" (UniqueName: \"kubernetes.io/projected/66392b25-1220-47fc-96d2-63e43a34ce70-kube-api-access-66r7p\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634117 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950f31f5-e378-4a18-96f7-04e2091e2ca3-serving-cert\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634133 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-888xr\" (UniqueName: \"kubernetes.io/projected/950f31f5-e378-4a18-96f7-04e2091e2ca3-kube-api-access-888xr\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634152 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-trusted-ca\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634233 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9038da7b-0b0e-44f6-abda-827d7de86d0d-audit-dir\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634559 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-oauth-serving-cert\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634582 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-audit-dir\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634598 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-audit-policies\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634614 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-stats-auth\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634633 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/642783e8-b8d6-4f1f-a313-a771a915ffb6-images\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634652 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-config\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634709 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn74z\" (UniqueName: \"kubernetes.io/projected/15119df3-c69c-4ee2-8a7a-f9373494178d-kube-api-access-dn74z\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634729 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-serving-cert\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634734 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-audit-dir\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634744 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634761 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634780 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-config\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634796 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-service-ca-bundle\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634858 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-console-config\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634875 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9tjh\" (UniqueName: \"kubernetes.io/projected/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-kube-api-access-v9tjh\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634889 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/642783e8-b8d6-4f1f-a313-a771a915ffb6-proxy-tls\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634915 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/df3e0f1c-d750-4b75-8f31-aa4faf37093c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.634985 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-trusted-ca-bundle\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635013 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-image-import-ca\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635029 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635080 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-config\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635102 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-client-ca\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635155 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635173 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-machine-approver-tls\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635192 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-metrics-certs\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635233 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95bf0de9-2c4f-4504-a083-49cda7af2c1a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635250 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356b0b5-8205-45e8-9ebb-0b2c7be780be-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635267 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-config\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635269 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/860ad6f7-00fc-4974-a703-389c3ca9d3a0-config\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635468 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r82lc\" (UniqueName: \"kubernetes.io/projected/09c2b041-5634-431c-bc61-5a4418c5296e-kube-api-access-r82lc\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635545 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z52m\" (UniqueName: \"kubernetes.io/projected/29a77e31-52d3-4f97-ad09-4422d5bd523a-kube-api-access-4z52m\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635626 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/642783e8-b8d6-4f1f-a313-a771a915ffb6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635645 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cecca3b-390d-4346-9bb4-586c9b7f56ea-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635665 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15119df3-c69c-4ee2-8a7a-f9373494178d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635681 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635697 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635713 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfmtv\" (UniqueName: \"kubernetes.io/projected/8cecca3b-390d-4346-9bb4-586c9b7f56ea-kube-api-access-lfmtv\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635732 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpcj8\" (UniqueName: \"kubernetes.io/projected/4de2ff5d-8c21-4412-b175-446ec4eec272-kube-api-access-vpcj8\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635746 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-client\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635766 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635784 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-ca\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635800 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356b0b5-8205-45e8-9ebb-0b2c7be780be-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635852 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635869 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-etcd-client\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635884 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-encryption-config\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635884 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635901 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.635951 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636033 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-config\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636056 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636094 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q28mq\" (UniqueName: \"kubernetes.io/projected/642783e8-b8d6-4f1f-a313-a771a915ffb6-kube-api-access-q28mq\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636110 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cecca3b-390d-4346-9bb4-586c9b7f56ea-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636127 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9038da7b-0b0e-44f6-abda-827d7de86d0d-node-pullsecrets\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636142 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-serving-cert\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636182 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-etcd-serving-ca\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636198 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15119df3-c69c-4ee2-8a7a-f9373494178d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636213 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29a77e31-52d3-4f97-ad09-4422d5bd523a-serving-cert\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636264 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcfxm\" (UniqueName: \"kubernetes.io/projected/846f52e4-6f0a-4969-9b1d-8465d3a2b730-kube-api-access-xcfxm\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636281 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636296 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-oauth-config\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636336 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df3e0f1c-d750-4b75-8f31-aa4faf37093c-serving-cert\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636359 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-service-ca\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636374 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr66j\" (UniqueName: \"kubernetes.io/projected/5356b0b5-8205-45e8-9ebb-0b2c7be780be-kube-api-access-rr66j\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636414 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e432ea72-cff7-4826-8001-125443ca158a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636423 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15119df3-c69c-4ee2-8a7a-f9373494178d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636433 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/846f52e4-6f0a-4969-9b1d-8465d3a2b730-service-ca-bundle\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.636647 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-config\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.637435 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-image-import-ca\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.637584 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.638211 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-client-ca\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.638754 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9038da7b-0b0e-44f6-abda-827d7de86d0d-etcd-serving-ca\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.639585 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-config\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.639661 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9038da7b-0b0e-44f6-abda-827d7de86d0d-node-pullsecrets\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.640732 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e432ea72-cff7-4826-8001-125443ca158a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.642084 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15119df3-c69c-4ee2-8a7a-f9373494178d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.645220 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-serving-cert\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.645265 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dckpn"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.645568 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.646337 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.647340 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.648331 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-c8jw7"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.649340 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-88r9w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.650415 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.651380 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.652367 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4zsg8"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.652469 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b38416c-b5e5-4195-b950-6e449e035c8b-serving-cert\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.653060 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-serving-cert\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.653293 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.653367 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-audit-policies\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.653423 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qxz55"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.653684 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950f31f5-e378-4a18-96f7-04e2091e2ca3-serving-cert\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.654117 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.654204 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.654375 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7pmwt"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.655064 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-encryption-config\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.655111 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-etcd-client\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.655455 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.655642 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.656467 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.657442 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gwtm7"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.657704 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9038da7b-0b0e-44f6-abda-827d7de86d0d-encryption-config\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.658522 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-66ndx"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.659738 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-srd6x"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.661108 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.662427 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7qwhs"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.663530 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.664825 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.665092 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.666047 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.667255 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-htg8d"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.668278 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.669137 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll54w"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.670539 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.671848 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.673001 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.673886 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.674356 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-plzvk"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.675416 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7pmwt"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.676562 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.677578 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qxz55"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.680521 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bd68z"] Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.681017 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.684569 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.705315 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.724737 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737239 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-serving-cert\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737286 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/055ff049-72f6-415e-afa6-b262b09be1da-config\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737321 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt6l2\" (UniqueName: \"kubernetes.io/projected/b6946c48-726b-4b01-a223-2a71d962694b-kube-api-access-rt6l2\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737345 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-service-ca-bundle\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737367 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9tjh\" (UniqueName: \"kubernetes.io/projected/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-kube-api-access-v9tjh\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737413 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/642783e8-b8d6-4f1f-a313-a771a915ffb6-proxy-tls\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737507 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737543 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6hp9\" (UniqueName: \"kubernetes.io/projected/3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c-kube-api-access-h6hp9\") pod \"migrator-59844c95c7-6bjfg\" (UID: \"3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737561 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-trusted-ca-bundle\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737580 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737600 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ec53082-404b-4679-89cf-84130cb32f60-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737618 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-machine-approver-tls\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737640 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737659 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-metrics-certs\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737676 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95bf0de9-2c4f-4504-a083-49cda7af2c1a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737694 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356b0b5-8205-45e8-9ebb-0b2c7be780be-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737711 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r82lc\" (UniqueName: \"kubernetes.io/projected/09c2b041-5634-431c-bc61-5a4418c5296e-kube-api-access-r82lc\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737728 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z52m\" (UniqueName: \"kubernetes.io/projected/29a77e31-52d3-4f97-ad09-4422d5bd523a-kube-api-access-4z52m\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737744 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/642783e8-b8d6-4f1f-a313-a771a915ffb6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737775 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737793 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737810 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737843 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-client\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737858 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7f7cc25b-f9a8-42a5-b114-929a466c16c9-profile-collector-cert\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737878 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356b0b5-8205-45e8-9ebb-0b2c7be780be-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737895 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/055ff049-72f6-415e-afa6-b262b09be1da-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737914 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737932 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737950 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q28mq\" (UniqueName: \"kubernetes.io/projected/642783e8-b8d6-4f1f-a313-a771a915ffb6-kube-api-access-q28mq\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737968 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cecca3b-390d-4346-9bb4-586c9b7f56ea-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.737985 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11c62d66-9fdc-480e-b259-dc768c6a04e5-secret-volume\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738012 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29a77e31-52d3-4f97-ad09-4422d5bd523a-serving-cert\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738028 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpflc\" (UniqueName: \"kubernetes.io/projected/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-kube-api-access-rpflc\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738048 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-oauth-config\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738076 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/846f52e4-6f0a-4969-9b1d-8465d3a2b730-service-ca-bundle\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738093 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-default-certificate\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738108 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95bf0de9-2c4f-4504-a083-49cda7af2c1a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738127 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-config\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738142 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7f7cc25b-f9a8-42a5-b114-929a466c16c9-srv-cert\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738158 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwtr7\" (UniqueName: \"kubernetes.io/projected/22165712-08e9-47cc-aaab-e5da9f92a0fb-kube-api-access-nwtr7\") pod \"control-plane-machine-set-operator-78cbb6b69f-pcrsf\" (UID: \"22165712-08e9-47cc-aaab-e5da9f92a0fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738176 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxp79\" (UniqueName: \"kubernetes.io/projected/2ec53082-404b-4679-89cf-84130cb32f60-kube-api-access-pxp79\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738198 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4de2ff5d-8c21-4412-b175-446ec4eec272-serving-cert\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738209 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-service-ca-bundle\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738216 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6gnl\" (UniqueName: \"kubernetes.io/projected/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-kube-api-access-p6gnl\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738261 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3a5d32-c068-4069-8f41-3bad3c040ac0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6v97j\" (UID: \"ca3a5d32-c068-4069-8f41-3bad3c040ac0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738283 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738309 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1fbf61d9-c3ca-4298-aa49-8f113925aec7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s6wmc\" (UID: \"1fbf61d9-c3ca-4298-aa49-8f113925aec7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738326 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5cdc38b1-0713-4b57-af9e-40071d064bc6-signing-cabundle\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738345 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/22142f3f-8cea-4691-b589-6f4c682740de-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-htg8d\" (UID: \"22142f3f-8cea-4691-b589-6f4c682740de\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738364 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sznc\" (UniqueName: \"kubernetes.io/projected/22142f3f-8cea-4691-b589-6f4c682740de-kube-api-access-5sznc\") pod \"multus-admission-controller-857f4d67dd-htg8d\" (UID: \"22142f3f-8cea-4691-b589-6f4c682740de\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738384 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738405 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22165712-08e9-47cc-aaab-e5da9f92a0fb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pcrsf\" (UID: \"22165712-08e9-47cc-aaab-e5da9f92a0fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738426 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738443 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738474 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nxc9\" (UniqueName: \"kubernetes.io/projected/eb898d9c-aa97-4aa4-a31f-ab8659120558-kube-api-access-2nxc9\") pod \"dns-operator-744455d44c-c8jw7\" (UID: \"eb898d9c-aa97-4aa4-a31f-ab8659120558\") " pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738507 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-serving-cert\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738522 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-service-ca\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738545 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eb898d9c-aa97-4aa4-a31f-ab8659120558-metrics-tls\") pod \"dns-operator-744455d44c-c8jw7\" (UID: \"eb898d9c-aa97-4aa4-a31f-ab8659120558\") " pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738563 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66r7p\" (UniqueName: \"kubernetes.io/projected/66392b25-1220-47fc-96d2-63e43a34ce70-kube-api-access-66r7p\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738579 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-oauth-serving-cert\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738605 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-proxy-tls\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738623 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e32f31a9-a83c-4935-8bb8-71ea757cb10f-tmpfs\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738647 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ec53082-404b-4679-89cf-84130cb32f60-trusted-ca\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738669 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738685 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738701 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-console-config\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738719 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5bld\" (UniqueName: \"kubernetes.io/projected/ca3a5d32-c068-4069-8f41-3bad3c040ac0-kube-api-access-q5bld\") pod \"package-server-manager-789f6589d5-6v97j\" (UID: \"ca3a5d32-c068-4069-8f41-3bad3c040ac0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738736 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvk75\" (UniqueName: \"kubernetes.io/projected/f760f393-f715-4e80-91ea-b8036cf6a3f8-kube-api-access-tvk75\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738754 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/df3e0f1c-d750-4b75-8f31-aa4faf37093c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738791 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rbfw\" (UniqueName: \"kubernetes.io/projected/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-kube-api-access-4rbfw\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738807 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f760f393-f715-4e80-91ea-b8036cf6a3f8-metrics-tls\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738838 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e32f31a9-a83c-4935-8bb8-71ea757cb10f-apiservice-cert\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738856 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/055ff049-72f6-415e-afa6-b262b09be1da-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738871 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5cdc38b1-0713-4b57-af9e-40071d064bc6-signing-key\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738890 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/860ad6f7-00fc-4974-a703-389c3ca9d3a0-config\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738907 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cecca3b-390d-4346-9bb4-586c9b7f56ea-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738928 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9csg\" (UniqueName: \"kubernetes.io/projected/5cdc38b1-0713-4b57-af9e-40071d064bc6-kube-api-access-j9csg\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738932 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356b0b5-8205-45e8-9ebb-0b2c7be780be-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.738985 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfmtv\" (UniqueName: \"kubernetes.io/projected/8cecca3b-390d-4346-9bb4-586c9b7f56ea-kube-api-access-lfmtv\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739005 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpcj8\" (UniqueName: \"kubernetes.io/projected/4de2ff5d-8c21-4412-b175-446ec4eec272-kube-api-access-vpcj8\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739042 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739063 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-ca\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739082 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ec53082-404b-4679-89cf-84130cb32f60-metrics-tls\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739120 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739147 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcfxm\" (UniqueName: \"kubernetes.io/projected/846f52e4-6f0a-4969-9b1d-8465d3a2b730-kube-api-access-xcfxm\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739163 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739199 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df3e0f1c-d750-4b75-8f31-aa4faf37093c-serving-cert\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739216 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-service-ca\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739232 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr66j\" (UniqueName: \"kubernetes.io/projected/5356b0b5-8205-45e8-9ebb-0b2c7be780be-kube-api-access-rr66j\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739248 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e32f31a9-a83c-4935-8bb8-71ea757cb10f-webhook-cert\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739290 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ff7n\" (UniqueName: \"kubernetes.io/projected/df3e0f1c-d750-4b75-8f31-aa4faf37093c-kube-api-access-8ff7n\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739308 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/860ad6f7-00fc-4974-a703-389c3ca9d3a0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739325 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739364 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhd5n\" (UniqueName: \"kubernetes.io/projected/f43b40c1-6db1-4026-a531-a020cb538423-kube-api-access-fhd5n\") pod \"downloads-7954f5f757-srd6x\" (UID: \"f43b40c1-6db1-4026-a531-a020cb538423\") " pod="openshift-console/downloads-7954f5f757-srd6x" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739385 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-config\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739394 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739401 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8cecca3b-390d-4346-9bb4-586c9b7f56ea-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739439 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x96fb\" (UniqueName: \"kubernetes.io/projected/11c62d66-9fdc-480e-b259-dc768c6a04e5-kube-api-access-x96fb\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.739604 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-trusted-ca-bundle\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.740011 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.740846 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-serving-cert\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.740906 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cecca3b-390d-4346-9bb4-586c9b7f56ea-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.741585 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-machine-approver-tls\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742420 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-auth-proxy-config\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742462 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-config\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742499 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09c2b041-5634-431c-bc61-5a4418c5296e-audit-dir\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742523 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95bf0de9-2c4f-4504-a083-49cda7af2c1a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742551 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742577 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f760f393-f715-4e80-91ea-b8036cf6a3f8-config-volume\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742602 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjtlm\" (UniqueName: \"kubernetes.io/projected/7f7cc25b-f9a8-42a5-b114-929a466c16c9-kube-api-access-fjtlm\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742960 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwt82\" (UniqueName: \"kubernetes.io/projected/1fbf61d9-c3ca-4298-aa49-8f113925aec7-kube-api-access-nwt82\") pod \"cluster-samples-operator-665b6dd947-s6wmc\" (UID: \"1fbf61d9-c3ca-4298-aa49-8f113925aec7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.742997 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/860ad6f7-00fc-4974-a703-389c3ca9d3a0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743034 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-trusted-ca\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743062 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhxts\" (UniqueName: \"kubernetes.io/projected/e32f31a9-a83c-4935-8bb8-71ea757cb10f-kube-api-access-rhxts\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743091 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-audit-policies\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743114 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-stats-auth\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743139 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/642783e8-b8d6-4f1f-a313-a771a915ffb6-images\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743161 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-srv-cert\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743186 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-config\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743210 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv6kn\" (UniqueName: \"kubernetes.io/projected/3854bccd-8788-4cce-8a17-8d06939f172b-kube-api-access-lv6kn\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.743773 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09c2b041-5634-431c-bc61-5a4418c5296e-audit-dir\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.744492 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356b0b5-8205-45e8-9ebb-0b2c7be780be-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.745901 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.746026 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-config\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.746061 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.746527 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/642783e8-b8d6-4f1f-a313-a771a915ffb6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.746543 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.746546 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.746806 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.747228 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-audit-policies\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.748120 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.748485 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-trusted-ca\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.748652 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-auth-proxy-config\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.750352 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/df3e0f1c-d750-4b75-8f31-aa4faf37093c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.749402 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.750396 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-config\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.749500 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-service-ca\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.749963 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-service-ca\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.750023 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-console-config\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.750304 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de2ff5d-8c21-4412-b175-446ec4eec272-config\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.749336 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-oauth-serving-cert\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.749402 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1fbf61d9-c3ca-4298-aa49-8f113925aec7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s6wmc\" (UID: \"1fbf61d9-c3ca-4298-aa49-8f113925aec7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.751656 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eb898d9c-aa97-4aa4-a31f-ab8659120558-metrics-tls\") pod \"dns-operator-744455d44c-c8jw7\" (UID: \"eb898d9c-aa97-4aa4-a31f-ab8659120558\") " pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.752434 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8cecca3b-390d-4346-9bb4-586c9b7f56ea-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.752683 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.752978 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4de2ff5d-8c21-4412-b175-446ec4eec272-serving-cert\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.753042 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29a77e31-52d3-4f97-ad09-4422d5bd523a-serving-cert\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.753196 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df3e0f1c-d750-4b75-8f31-aa4faf37093c-serving-cert\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.753461 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-serving-cert\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.753471 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-client\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.753664 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-oauth-config\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.753723 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.754285 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.754288 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.764853 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.773309 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95bf0de9-2c4f-4504-a083-49cda7af2c1a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.785100 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.793262 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95bf0de9-2c4f-4504-a083-49cda7af2c1a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.805114 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.825956 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.835945 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-config\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.845358 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.845728 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwtr7\" (UniqueName: \"kubernetes.io/projected/22165712-08e9-47cc-aaab-e5da9f92a0fb-kube-api-access-nwtr7\") pod \"control-plane-machine-set-operator-78cbb6b69f-pcrsf\" (UID: \"22165712-08e9-47cc-aaab-e5da9f92a0fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.846062 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxp79\" (UniqueName: \"kubernetes.io/projected/2ec53082-404b-4679-89cf-84130cb32f60-kube-api-access-pxp79\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.846184 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7f7cc25b-f9a8-42a5-b114-929a466c16c9-srv-cert\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.846446 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.846763 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3a5d32-c068-4069-8f41-3bad3c040ac0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6v97j\" (UID: \"ca3a5d32-c068-4069-8f41-3bad3c040ac0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847032 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/22142f3f-8cea-4691-b589-6f4c682740de-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-htg8d\" (UID: \"22142f3f-8cea-4691-b589-6f4c682740de\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847200 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sznc\" (UniqueName: \"kubernetes.io/projected/22142f3f-8cea-4691-b589-6f4c682740de-kube-api-access-5sznc\") pod \"multus-admission-controller-857f4d67dd-htg8d\" (UID: \"22142f3f-8cea-4691-b589-6f4c682740de\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847268 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5cdc38b1-0713-4b57-af9e-40071d064bc6-signing-cabundle\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847330 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22165712-08e9-47cc-aaab-e5da9f92a0fb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pcrsf\" (UID: \"22165712-08e9-47cc-aaab-e5da9f92a0fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847366 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847468 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-proxy-tls\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847509 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e32f31a9-a83c-4935-8bb8-71ea757cb10f-tmpfs\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847541 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ec53082-404b-4679-89cf-84130cb32f60-trusted-ca\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847604 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvk75\" (UniqueName: \"kubernetes.io/projected/f760f393-f715-4e80-91ea-b8036cf6a3f8-kube-api-access-tvk75\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847640 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5bld\" (UniqueName: \"kubernetes.io/projected/ca3a5d32-c068-4069-8f41-3bad3c040ac0-kube-api-access-q5bld\") pod \"package-server-manager-789f6589d5-6v97j\" (UID: \"ca3a5d32-c068-4069-8f41-3bad3c040ac0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847671 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f760f393-f715-4e80-91ea-b8036cf6a3f8-metrics-tls\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847709 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rbfw\" (UniqueName: \"kubernetes.io/projected/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-kube-api-access-4rbfw\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847740 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e32f31a9-a83c-4935-8bb8-71ea757cb10f-apiservice-cert\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847772 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5cdc38b1-0713-4b57-af9e-40071d064bc6-signing-key\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847808 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/055ff049-72f6-415e-afa6-b262b09be1da-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847879 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9csg\" (UniqueName: \"kubernetes.io/projected/5cdc38b1-0713-4b57-af9e-40071d064bc6-kube-api-access-j9csg\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.847960 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ec53082-404b-4679-89cf-84130cb32f60-metrics-tls\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848016 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848060 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e32f31a9-a83c-4935-8bb8-71ea757cb10f-webhook-cert\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848115 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e32f31a9-a83c-4935-8bb8-71ea757cb10f-tmpfs\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848128 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhd5n\" (UniqueName: \"kubernetes.io/projected/f43b40c1-6db1-4026-a531-a020cb538423-kube-api-access-fhd5n\") pod \"downloads-7954f5f757-srd6x\" (UID: \"f43b40c1-6db1-4026-a531-a020cb538423\") " pod="openshift-console/downloads-7954f5f757-srd6x" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848225 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x96fb\" (UniqueName: \"kubernetes.io/projected/11c62d66-9fdc-480e-b259-dc768c6a04e5-kube-api-access-x96fb\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848273 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848341 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f760f393-f715-4e80-91ea-b8036cf6a3f8-config-volume\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848393 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjtlm\" (UniqueName: \"kubernetes.io/projected/7f7cc25b-f9a8-42a5-b114-929a466c16c9-kube-api-access-fjtlm\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848443 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhxts\" (UniqueName: \"kubernetes.io/projected/e32f31a9-a83c-4935-8bb8-71ea757cb10f-kube-api-access-rhxts\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848508 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-srv-cert\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848549 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv6kn\" (UniqueName: \"kubernetes.io/projected/3854bccd-8788-4cce-8a17-8d06939f172b-kube-api-access-lv6kn\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848580 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/055ff049-72f6-415e-afa6-b262b09be1da-config\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848612 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt6l2\" (UniqueName: \"kubernetes.io/projected/b6946c48-726b-4b01-a223-2a71d962694b-kube-api-access-rt6l2\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848664 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848717 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6hp9\" (UniqueName: \"kubernetes.io/projected/3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c-kube-api-access-h6hp9\") pod \"migrator-59844c95c7-6bjfg\" (UID: \"3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.848793 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.849238 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ec53082-404b-4679-89cf-84130cb32f60-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.849427 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.849509 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7f7cc25b-f9a8-42a5-b114-929a466c16c9-profile-collector-cert\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.849547 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/055ff049-72f6-415e-afa6-b262b09be1da-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.849581 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11c62d66-9fdc-480e-b259-dc768c6a04e5-secret-volume\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.849626 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpflc\" (UniqueName: \"kubernetes.io/projected/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-kube-api-access-rpflc\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.849735 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.865113 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.879177 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/860ad6f7-00fc-4974-a703-389c3ca9d3a0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.885266 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.894205 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/860ad6f7-00fc-4974-a703-389c3ca9d3a0-config\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.905654 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.925007 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.933184 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/29a77e31-52d3-4f97-ad09-4422d5bd523a-etcd-ca\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.949907 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.954110 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/846f52e4-6f0a-4969-9b1d-8465d3a2b730-service-ca-bundle\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.965572 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.977009 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-default-certificate\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:46 crc kubenswrapper[4838]: I1207 09:08:46.985758 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.000338 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-stats-auth\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.005596 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.014586 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/846f52e4-6f0a-4969-9b1d-8465d3a2b730-metrics-certs\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.025677 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.046129 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.066129 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.086040 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.097432 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/642783e8-b8d6-4f1f-a313-a771a915ffb6-images\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.106348 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.125885 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.132170 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/642783e8-b8d6-4f1f-a313-a771a915ffb6-proxy-tls\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.145791 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.153717 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f760f393-f715-4e80-91ea-b8036cf6a3f8-metrics-tls\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.165845 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.185733 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.196695 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f760f393-f715-4e80-91ea-b8036cf6a3f8-config-volume\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.206366 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.214061 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3a5d32-c068-4069-8f41-3bad3c040ac0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6v97j\" (UID: \"ca3a5d32-c068-4069-8f41-3bad3c040ac0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.226138 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.246188 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.252658 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e32f31a9-a83c-4935-8bb8-71ea757cb10f-apiservice-cert\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.253600 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e32f31a9-a83c-4935-8bb8-71ea757cb10f-webhook-cert\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.265915 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.271787 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5cdc38b1-0713-4b57-af9e-40071d064bc6-signing-key\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.285678 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.305097 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.325642 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.345257 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.365523 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.386371 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.405720 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.426371 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.429531 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5cdc38b1-0713-4b57-af9e-40071d064bc6-signing-cabundle\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.445679 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.468743 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.485318 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.505523 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.525197 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.545751 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.565575 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.572134 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ec53082-404b-4679-89cf-84130cb32f60-metrics-tls\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.584219 4838 request.go:700] Waited for 1.0010301s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.585910 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.606282 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.639407 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.649850 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ec53082-404b-4679-89cf-84130cb32f60-trusted-ca\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.668145 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.687423 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.706057 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.712909 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-proxy-tls\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.724887 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.733142 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/22142f3f-8cea-4691-b589-6f4c682740de-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-htg8d\" (UID: \"22142f3f-8cea-4691-b589-6f4c682740de\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.744878 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.764932 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.773865 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/055ff049-72f6-415e-afa6-b262b09be1da-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.786649 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.789675 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/055ff049-72f6-415e-afa6-b262b09be1da-config\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.805886 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.812422 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.814134 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11c62d66-9fdc-480e-b259-dc768c6a04e5-secret-volume\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.815025 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7f7cc25b-f9a8-42a5-b114-929a466c16c9-profile-collector-cert\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.825425 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.845954 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.847509 4838 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.847575 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22165712-08e9-47cc-aaab-e5da9f92a0fb-control-plane-machine-set-operator-tls podName:22165712-08e9-47cc-aaab-e5da9f92a0fb nodeName:}" failed. No retries permitted until 2025-12-07 09:08:48.347553636 +0000 UTC m=+145.054872663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/22165712-08e9-47cc-aaab-e5da9f92a0fb-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-pcrsf" (UID: "22165712-08e9-47cc-aaab-e5da9f92a0fb") : failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.847620 4838 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.847719 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca podName:3854bccd-8788-4cce-8a17-8d06939f172b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:48.34769244 +0000 UTC m=+145.055011497 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca") pod "marketplace-operator-79b997595-ll54w" (UID: "3854bccd-8788-4cce-8a17-8d06939f172b") : failed to sync configmap cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.848342 4838 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.849894 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics podName:3854bccd-8788-4cce-8a17-8d06939f172b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:48.349871269 +0000 UTC m=+145.057190316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics") pod "marketplace-operator-79b997595-ll54w" (UID: "3854bccd-8788-4cce-8a17-8d06939f172b") : failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.848714 4838 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.849960 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume podName:11c62d66-9fdc-480e-b259-dc768c6a04e5 nodeName:}" failed. No retries permitted until 2025-12-07 09:08:48.349947701 +0000 UTC m=+145.057266758 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume") pod "collect-profiles-29418300-jr5mm" (UID: "11c62d66-9fdc-480e-b259-dc768c6a04e5") : failed to sync configmap cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.848740 4838 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.850008 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-srv-cert podName:b6946c48-726b-4b01-a223-2a71d962694b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:48.349997562 +0000 UTC m=+145.057316619 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-srv-cert") pod "olm-operator-6b444d44fb-6rfh7" (UID: "b6946c48-726b-4b01-a223-2a71d962694b") : failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.849471 4838 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.850063 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-serving-cert podName:b2067c5a-8073-4b7a-b929-5a4b5ff6846f nodeName:}" failed. No retries permitted until 2025-12-07 09:08:48.350053174 +0000 UTC m=+145.057372231 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-tpxz5" (UID: "b2067c5a-8073-4b7a-b929-5a4b5ff6846f") : failed to sync secret cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.850137 4838 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: E1207 09:08:47.850237 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-config podName:b2067c5a-8073-4b7a-b929-5a4b5ff6846f nodeName:}" failed. No retries permitted until 2025-12-07 09:08:48.350215488 +0000 UTC m=+145.057534615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-config") pod "kube-storage-version-migrator-operator-b67b599dd-tpxz5" (UID: "b2067c5a-8073-4b7a-b929-5a4b5ff6846f") : failed to sync configmap cache: timed out waiting for the condition Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.852035 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7f7cc25b-f9a8-42a5-b114-929a466c16c9-srv-cert\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.894002 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvxgm\" (UniqueName: \"kubernetes.io/projected/9b38416c-b5e5-4195-b950-6e449e035c8b-kube-api-access-cvxgm\") pod \"route-controller-manager-6576b87f9c-ccfmf\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.905239 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.912132 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kd4z\" (UniqueName: \"kubernetes.io/projected/9038da7b-0b0e-44f6-abda-827d7de86d0d-kube-api-access-8kd4z\") pod \"apiserver-76f77b778f-ng7p6\" (UID: \"9038da7b-0b0e-44f6-abda-827d7de86d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.915213 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.925679 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.946094 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.955211 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.965948 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 07 09:08:47 crc kubenswrapper[4838]: I1207 09:08:47.997607 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.006936 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.026249 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.048956 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.068999 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.087992 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.105790 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.126346 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.145744 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.170021 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.179051 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf"] Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.186090 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.187152 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ng7p6"] Dec 07 09:08:48 crc kubenswrapper[4838]: W1207 09:08:48.193946 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9038da7b_0b0e_44f6_abda_827d7de86d0d.slice/crio-ea19eab5c2b82f83180dfc4fbadefa5693d5fd5fa21911b03e56c93e976f1e73 WatchSource:0}: Error finding container ea19eab5c2b82f83180dfc4fbadefa5693d5fd5fa21911b03e56c93e976f1e73: Status 404 returned error can't find the container with id ea19eab5c2b82f83180dfc4fbadefa5693d5fd5fa21911b03e56c93e976f1e73 Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.205087 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.225148 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.244755 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.265596 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.304389 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzhml\" (UniqueName: \"kubernetes.io/projected/e432ea72-cff7-4826-8001-125443ca158a-kube-api-access-mzhml\") pod \"machine-api-operator-5694c8668f-qll5w\" (UID: \"e432ea72-cff7-4826-8001-125443ca158a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.327984 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq68w\" (UniqueName: \"kubernetes.io/projected/c34b4216-fa9a-4fd9-9c00-53e62f4c73a9-kube-api-access-zq68w\") pod \"apiserver-7bbb656c7d-ggzfg\" (UID: \"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.345776 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.356394 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" event={"ID":"9038da7b-0b0e-44f6-abda-827d7de86d0d","Type":"ContainerStarted","Data":"ea19eab5c2b82f83180dfc4fbadefa5693d5fd5fa21911b03e56c93e976f1e73"} Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.357615 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" event={"ID":"9b38416c-b5e5-4195-b950-6e449e035c8b","Type":"ContainerStarted","Data":"cd9ede871ca01861ca4285d84d408fab645da056f33270c89a0ddefba73d7ca1"} Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.378429 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.378547 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.378608 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-srv-cert\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.378652 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.378716 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.378807 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22165712-08e9-47cc-aaab-e5da9f92a0fb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pcrsf\" (UID: \"22165712-08e9-47cc-aaab-e5da9f92a0fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.378855 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.379376 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.380792 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.381644 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn74z\" (UniqueName: \"kubernetes.io/projected/15119df3-c69c-4ee2-8a7a-f9373494178d-kube-api-access-dn74z\") pod \"openshift-apiserver-operator-796bbdcf4f-78x4g\" (UID: \"15119df3-c69c-4ee2-8a7a-f9373494178d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.381741 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.382707 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6946c48-726b-4b01-a223-2a71d962694b-srv-cert\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.383685 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.384393 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.385506 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22165712-08e9-47cc-aaab-e5da9f92a0fb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pcrsf\" (UID: \"22165712-08e9-47cc-aaab-e5da9f92a0fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.396724 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.405427 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.425544 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.446739 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.472216 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.485550 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.493849 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-888xr\" (UniqueName: \"kubernetes.io/projected/950f31f5-e378-4a18-96f7-04e2091e2ca3-kube-api-access-888xr\") pod \"controller-manager-879f6c89f-wmz9f\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.506665 4838 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.527171 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.545584 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.565837 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.570100 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.585363 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qll5w"] Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.585712 4838 request.go:700] Waited for 1.904501556s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.587376 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 07 09:08:48 crc kubenswrapper[4838]: W1207 09:08:48.595269 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode432ea72_cff7_4826_8001_125443ca158a.slice/crio-b8d85434bb45961936e0c520e88de6295e75c7fd7172e42eac7273582b70a83a WatchSource:0}: Error finding container b8d85434bb45961936e0c520e88de6295e75c7fd7172e42eac7273582b70a83a: Status 404 returned error can't find the container with id b8d85434bb45961936e0c520e88de6295e75c7fd7172e42eac7273582b70a83a Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.624930 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9tjh\" (UniqueName: \"kubernetes.io/projected/03b837e6-4cd2-43b3-b6b4-a8908c7fbab4-kube-api-access-v9tjh\") pod \"machine-approver-56656f9798-rgbcf\" (UID: \"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.642491 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6gnl\" (UniqueName: \"kubernetes.io/projected/21433b7b-de00-4a67-8b98-ac5f7ce89ad8-kube-api-access-p6gnl\") pod \"console-operator-58897d9998-4zsg8\" (UID: \"21433b7b-de00-4a67-8b98-ac5f7ce89ad8\") " pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.649512 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.668125 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r82lc\" (UniqueName: \"kubernetes.io/projected/09c2b041-5634-431c-bc61-5a4418c5296e-kube-api-access-r82lc\") pod \"oauth-openshift-558db77b4-t4fqk\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.671263 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.687568 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q28mq\" (UniqueName: \"kubernetes.io/projected/642783e8-b8d6-4f1f-a313-a771a915ffb6-kube-api-access-q28mq\") pod \"machine-config-operator-74547568cd-k9jbv\" (UID: \"642783e8-b8d6-4f1f-a313-a771a915ffb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.704632 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpcj8\" (UniqueName: \"kubernetes.io/projected/4de2ff5d-8c21-4412-b175-446ec4eec272-kube-api-access-vpcj8\") pod \"authentication-operator-69f744f599-dckpn\" (UID: \"4de2ff5d-8c21-4412-b175-446ec4eec272\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.721084 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr66j\" (UniqueName: \"kubernetes.io/projected/5356b0b5-8205-45e8-9ebb-0b2c7be780be-kube-api-access-rr66j\") pod \"openshift-controller-manager-operator-756b6f6bc6-cmsvz\" (UID: \"5356b0b5-8205-45e8-9ebb-0b2c7be780be\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.742926 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ff7n\" (UniqueName: \"kubernetes.io/projected/df3e0f1c-d750-4b75-8f31-aa4faf37093c-kube-api-access-8ff7n\") pod \"openshift-config-operator-7777fb866f-sdc4k\" (UID: \"df3e0f1c-d750-4b75-8f31-aa4faf37093c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.758160 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/860ad6f7-00fc-4974-a703-389c3ca9d3a0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bdbx5\" (UID: \"860ad6f7-00fc-4974-a703-389c3ca9d3a0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.771900 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.779697 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg"] Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.784840 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcfxm\" (UniqueName: \"kubernetes.io/projected/846f52e4-6f0a-4969-9b1d-8465d3a2b730-kube-api-access-xcfxm\") pod \"router-default-5444994796-pvp5p\" (UID: \"846f52e4-6f0a-4969-9b1d-8465d3a2b730\") " pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.788249 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.807074 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nxc9\" (UniqueName: \"kubernetes.io/projected/eb898d9c-aa97-4aa4-a31f-ab8659120558-kube-api-access-2nxc9\") pod \"dns-operator-744455d44c-c8jw7\" (UID: \"eb898d9c-aa97-4aa4-a31f-ab8659120558\") " pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.811588 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.825974 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.841488 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cecca3b-390d-4346-9bb4-586c9b7f56ea-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.841672 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.852472 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwt82\" (UniqueName: \"kubernetes.io/projected/1fbf61d9-c3ca-4298-aa49-8f113925aec7-kube-api-access-nwt82\") pod \"cluster-samples-operator-665b6dd947-s6wmc\" (UID: \"1fbf61d9-c3ca-4298-aa49-8f113925aec7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.859044 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.867342 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.867418 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.870575 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95bf0de9-2c4f-4504-a083-49cda7af2c1a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jdk8t\" (UID: \"95bf0de9-2c4f-4504-a083-49cda7af2c1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.876160 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.884925 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.891616 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z52m\" (UniqueName: \"kubernetes.io/projected/29a77e31-52d3-4f97-ad09-4422d5bd523a-kube-api-access-4z52m\") pod \"etcd-operator-b45778765-88r9w\" (UID: \"29a77e31-52d3-4f97-ad09-4422d5bd523a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.915305 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfmtv\" (UniqueName: \"kubernetes.io/projected/8cecca3b-390d-4346-9bb4-586c9b7f56ea-kube-api-access-lfmtv\") pod \"cluster-image-registry-operator-dc59b4c8b-mjwmq\" (UID: \"8cecca3b-390d-4346-9bb4-586c9b7f56ea\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.932305 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66r7p\" (UniqueName: \"kubernetes.io/projected/66392b25-1220-47fc-96d2-63e43a34ce70-kube-api-access-66r7p\") pod \"console-f9d7485db-xrdpm\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.943148 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwtr7\" (UniqueName: \"kubernetes.io/projected/22165712-08e9-47cc-aaab-e5da9f92a0fb-kube-api-access-nwtr7\") pod \"control-plane-machine-set-operator-78cbb6b69f-pcrsf\" (UID: \"22165712-08e9-47cc-aaab-e5da9f92a0fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.969268 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxp79\" (UniqueName: \"kubernetes.io/projected/2ec53082-404b-4679-89cf-84130cb32f60-kube-api-access-pxp79\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.976164 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wmz9f"] Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.987763 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sznc\" (UniqueName: \"kubernetes.io/projected/22142f3f-8cea-4691-b589-6f4c682740de-kube-api-access-5sznc\") pod \"multus-admission-controller-857f4d67dd-htg8d\" (UID: \"22142f3f-8cea-4691-b589-6f4c682740de\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:48 crc kubenswrapper[4838]: I1207 09:08:48.989196 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.008853 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvk75\" (UniqueName: \"kubernetes.io/projected/f760f393-f715-4e80-91ea-b8036cf6a3f8-kube-api-access-tvk75\") pod \"dns-default-7qwhs\" (UID: \"f760f393-f715-4e80-91ea-b8036cf6a3f8\") " pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.026469 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5bld\" (UniqueName: \"kubernetes.io/projected/ca3a5d32-c068-4069-8f41-3bad3c040ac0-kube-api-access-q5bld\") pod \"package-server-manager-789f6589d5-6v97j\" (UID: \"ca3a5d32-c068-4069-8f41-3bad3c040ac0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.041397 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rbfw\" (UniqueName: \"kubernetes.io/projected/b2067c5a-8073-4b7a-b929-5a4b5ff6846f-kube-api-access-4rbfw\") pod \"kube-storage-version-migrator-operator-b67b599dd-tpxz5\" (UID: \"b2067c5a-8073-4b7a-b929-5a4b5ff6846f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.045360 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g"] Dec 07 09:08:49 crc kubenswrapper[4838]: W1207 09:08:49.062496 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod950f31f5_e378_4a18_96f7_04e2091e2ca3.slice/crio-ba28e2fa50928e7e957e63da3cf9cd16552fb5b96f412d3f7241555ec42d1ad9 WatchSource:0}: Error finding container ba28e2fa50928e7e957e63da3cf9cd16552fb5b96f412d3f7241555ec42d1ad9: Status 404 returned error can't find the container with id ba28e2fa50928e7e957e63da3cf9cd16552fb5b96f412d3f7241555ec42d1ad9 Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.077881 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/055ff049-72f6-415e-afa6-b262b09be1da-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vtrtm\" (UID: \"055ff049-72f6-415e-afa6-b262b09be1da\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.080830 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.084756 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9csg\" (UniqueName: \"kubernetes.io/projected/5cdc38b1-0713-4b57-af9e-40071d064bc6-kube-api-access-j9csg\") pod \"service-ca-9c57cc56f-gwtm7\" (UID: \"5cdc38b1-0713-4b57-af9e-40071d064bc6\") " pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.095794 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.102915 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.108470 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhd5n\" (UniqueName: \"kubernetes.io/projected/f43b40c1-6db1-4026-a531-a020cb538423-kube-api-access-fhd5n\") pod \"downloads-7954f5f757-srd6x\" (UID: \"f43b40c1-6db1-4026-a531-a020cb538423\") " pod="openshift-console/downloads-7954f5f757-srd6x" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.119464 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.134775 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4fqk"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.139584 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x96fb\" (UniqueName: \"kubernetes.io/projected/11c62d66-9fdc-480e-b259-dc768c6a04e5-kube-api-access-x96fb\") pod \"collect-profiles-29418300-jr5mm\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.141202 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.146547 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjtlm\" (UniqueName: \"kubernetes.io/projected/7f7cc25b-f9a8-42a5-b114-929a466c16c9-kube-api-access-fjtlm\") pod \"catalog-operator-68c6474976-p246s\" (UID: \"7f7cc25b-f9a8-42a5-b114-929a466c16c9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.183593 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhxts\" (UniqueName: \"kubernetes.io/projected/e32f31a9-a83c-4935-8bb8-71ea757cb10f-kube-api-access-rhxts\") pod \"packageserver-d55dfcdfc-lcg8w\" (UID: \"e32f31a9-a83c-4935-8bb8-71ea757cb10f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.193465 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.199019 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt6l2\" (UniqueName: \"kubernetes.io/projected/b6946c48-726b-4b01-a223-2a71d962694b-kube-api-access-rt6l2\") pod \"olm-operator-6b444d44fb-6rfh7\" (UID: \"b6946c48-726b-4b01-a223-2a71d962694b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.202092 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.209709 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.217538 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv6kn\" (UniqueName: \"kubernetes.io/projected/3854bccd-8788-4cce-8a17-8d06939f172b-kube-api-access-lv6kn\") pod \"marketplace-operator-79b997595-ll54w\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.220406 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.230161 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-srd6x" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.231993 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6hp9\" (UniqueName: \"kubernetes.io/projected/3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c-kube-api-access-h6hp9\") pod \"migrator-59844c95c7-6bjfg\" (UID: \"3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.258251 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.267869 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ec53082-404b-4679-89cf-84130cb32f60-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rzx2t\" (UID: \"2ec53082-404b-4679-89cf-84130cb32f60\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.267882 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.268029 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpflc\" (UniqueName: \"kubernetes.io/projected/94a2d3db-cde7-4bad-be14-fe5bb0d3e450-kube-api-access-rpflc\") pod \"machine-config-controller-84d6567774-zsdrc\" (UID: \"94a2d3db-cde7-4bad-be14-fe5bb0d3e450\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.270666 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4zsg8"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.270901 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.283022 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.297618 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-certificates\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.297654 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-bound-sa-token\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.297680 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.297698 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.297761 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.297785 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-trusted-ca\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.299177 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:49.799159017 +0000 UTC m=+146.506478034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.299941 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.300271 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21071ba2-96ab-4eab-b2d6-92264e094685-config\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.300353 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvhmh\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-kube-api-access-bvhmh\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.300387 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-tls\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.300418 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf2fj\" (UniqueName: \"kubernetes.io/projected/21071ba2-96ab-4eab-b2d6-92264e094685-kube-api-access-kf2fj\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.300458 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21071ba2-96ab-4eab-b2d6-92264e094685-serving-cert\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.315273 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dckpn"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.334217 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.334592 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.336723 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-c8jw7"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.336822 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.412250 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.412337 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:49.912318074 +0000 UTC m=+146.619637091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.417653 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/13435500-f37a-4b8d-a396-1714f365cbc6-node-bootstrap-token\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.423656 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-certificates\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.425403 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-certificates\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.425546 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-bound-sa-token\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.428597 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" event={"ID":"15119df3-c69c-4ee2-8a7a-f9373494178d","Type":"ContainerStarted","Data":"2977ff22bb5ab7c3e8880fd3252c09040bf647329d8f4450053bbf9f9b98ef9e"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.429083 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm5n6\" (UniqueName: \"kubernetes.io/projected/c5adfada-3921-46ab-8be3-b9e7e812d276-kube-api-access-wm5n6\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.432996 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.433418 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.434140 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9728r\" (UniqueName: \"kubernetes.io/projected/13435500-f37a-4b8d-a396-1714f365cbc6-kube-api-access-9728r\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.434431 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.434894 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/13435500-f37a-4b8d-a396-1714f365cbc6-certs\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435154 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435273 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-mountpoint-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435441 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-trusted-ca\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435531 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21071ba2-96ab-4eab-b2d6-92264e094685-config\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435629 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-csi-data-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435712 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvhmh\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-kube-api-access-bvhmh\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435792 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-tls\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.435912 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-registration-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.436034 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf2fj\" (UniqueName: \"kubernetes.io/projected/21071ba2-96ab-4eab-b2d6-92264e094685-kube-api-access-kf2fj\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.436124 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-socket-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.436280 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtsk2\" (UniqueName: \"kubernetes.io/projected/9bf3b45c-616f-465f-8619-05d4690d6857-kube-api-access-jtsk2\") pod \"ingress-canary-qxz55\" (UID: \"9bf3b45c-616f-465f-8619-05d4690d6857\") " pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.436326 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21071ba2-96ab-4eab-b2d6-92264e094685-serving-cert\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.436374 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-plugins-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.436392 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9bf3b45c-616f-465f-8619-05d4690d6857-cert\") pod \"ingress-canary-qxz55\" (UID: \"9bf3b45c-616f-465f-8619-05d4690d6857\") " pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.456337 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-trusted-ca\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.460326 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:49.960305729 +0000 UTC m=+146.667624746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.460367 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.461776 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21071ba2-96ab-4eab-b2d6-92264e094685-config\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: W1207 09:08:49.471388 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21433b7b_de00_4a67_8b98_ac5f7ce89ad8.slice/crio-4bf780fe52c66d2eda5601cd2d3ab5d38ac3a562efe7ee77c8328c6d6c3c17f6 WatchSource:0}: Error finding container 4bf780fe52c66d2eda5601cd2d3ab5d38ac3a562efe7ee77c8328c6d6c3c17f6: Status 404 returned error can't find the container with id 4bf780fe52c66d2eda5601cd2d3ab5d38ac3a562efe7ee77c8328c6d6c3c17f6 Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.479780 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-bound-sa-token\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.480542 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" event={"ID":"950f31f5-e378-4a18-96f7-04e2091e2ca3","Type":"ContainerStarted","Data":"775c31257668362f6bc0212b222993cdc1d81972052e5d369c5269c5df450c75"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.480581 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" event={"ID":"950f31f5-e378-4a18-96f7-04e2091e2ca3","Type":"ContainerStarted","Data":"ba28e2fa50928e7e957e63da3cf9cd16552fb5b96f412d3f7241555ec42d1ad9"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.480600 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-tls\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.481314 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.485879 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21071ba2-96ab-4eab-b2d6-92264e094685-serving-cert\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.486435 4838 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-wmz9f container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.486469 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" podUID="950f31f5-e378-4a18-96f7-04e2091e2ca3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.491296 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf2fj\" (UniqueName: \"kubernetes.io/projected/21071ba2-96ab-4eab-b2d6-92264e094685-kube-api-access-kf2fj\") pod \"service-ca-operator-777779d784-plzvk\" (UID: \"21071ba2-96ab-4eab-b2d6-92264e094685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.506758 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvhmh\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-kube-api-access-bvhmh\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.513994 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.520140 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.521228 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537177 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.537345 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.037322003 +0000 UTC m=+146.744641020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537368 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-csi-data-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537442 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-registration-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537464 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-socket-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537484 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtsk2\" (UniqueName: \"kubernetes.io/projected/9bf3b45c-616f-465f-8619-05d4690d6857-kube-api-access-jtsk2\") pod \"ingress-canary-qxz55\" (UID: \"9bf3b45c-616f-465f-8619-05d4690d6857\") " pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537503 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-plugins-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537518 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9bf3b45c-616f-465f-8619-05d4690d6857-cert\") pod \"ingress-canary-qxz55\" (UID: \"9bf3b45c-616f-465f-8619-05d4690d6857\") " pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537547 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/13435500-f37a-4b8d-a396-1714f365cbc6-node-bootstrap-token\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537609 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm5n6\" (UniqueName: \"kubernetes.io/projected/c5adfada-3921-46ab-8be3-b9e7e812d276-kube-api-access-wm5n6\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537651 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9728r\" (UniqueName: \"kubernetes.io/projected/13435500-f37a-4b8d-a396-1714f365cbc6-kube-api-access-9728r\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537687 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/13435500-f37a-4b8d-a396-1714f365cbc6-certs\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537719 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537738 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-mountpoint-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537833 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-mountpoint-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.537899 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-csi-data-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.538739 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-registration-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.538786 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-socket-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.541475 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c5adfada-3921-46ab-8be3-b9e7e812d276-plugins-dir\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.544508 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.044495108 +0000 UTC m=+146.751814125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.544888 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.545485 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" event={"ID":"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4","Type":"ContainerStarted","Data":"ab10fe5c03c69fae42e9cfd5c2fa9a521e3d64abb8c07467bea5c17d5c9330ca"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.547794 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/13435500-f37a-4b8d-a396-1714f365cbc6-node-bootstrap-token\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.549156 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/13435500-f37a-4b8d-a396-1714f365cbc6-certs\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.553601 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9bf3b45c-616f-465f-8619-05d4690d6857-cert\") pod \"ingress-canary-qxz55\" (UID: \"9bf3b45c-616f-465f-8619-05d4690d6857\") " pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:49 crc kubenswrapper[4838]: W1207 09:08:49.555283 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4de2ff5d_8c21_4412_b175_446ec4eec272.slice/crio-ffcd637ac9e5c89cd5c8bec8206cecf362df4eb7c3a14b1ebdae71650e5837c7 WatchSource:0}: Error finding container ffcd637ac9e5c89cd5c8bec8206cecf362df4eb7c3a14b1ebdae71650e5837c7: Status 404 returned error can't find the container with id ffcd637ac9e5c89cd5c8bec8206cecf362df4eb7c3a14b1ebdae71650e5837c7 Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.557947 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" event={"ID":"09c2b041-5634-431c-bc61-5a4418c5296e","Type":"ContainerStarted","Data":"87680cf1c7378825fac8b7a94f57090bf4aeef655ce18f9da77a928e3bbede09"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.573120 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pvp5p" event={"ID":"846f52e4-6f0a-4969-9b1d-8465d3a2b730","Type":"ContainerStarted","Data":"9c1a86d200027a1a65e3affaf4fe5ed61a3dafb24cbe62e3387157a44889dde5"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.580386 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" event={"ID":"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9","Type":"ContainerStarted","Data":"769b336afaf99c1eda5c7899a1d5bfd04bbe01da67e78d845459172ffd4d3d3d"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.582546 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtsk2\" (UniqueName: \"kubernetes.io/projected/9bf3b45c-616f-465f-8619-05d4690d6857-kube-api-access-jtsk2\") pod \"ingress-canary-qxz55\" (UID: \"9bf3b45c-616f-465f-8619-05d4690d6857\") " pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.583974 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" event={"ID":"9b38416c-b5e5-4195-b950-6e449e035c8b","Type":"ContainerStarted","Data":"a4f1cd011880020db373ceb43c25b7c5d1479ad3b2fb01b0f19b321701ff471e"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.585964 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.595658 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.606928 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.608396 4838 generic.go:334] "Generic (PLEG): container finished" podID="9038da7b-0b0e-44f6-abda-827d7de86d0d" containerID="ddec23ed56ec799eadb75595d6d96dc2a9d6e55f2c29acd19ba2967c001ad231" exitCode=0 Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.608459 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" event={"ID":"9038da7b-0b0e-44f6-abda-827d7de86d0d","Type":"ContainerDied","Data":"ddec23ed56ec799eadb75595d6d96dc2a9d6e55f2c29acd19ba2967c001ad231"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.615205 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm5n6\" (UniqueName: \"kubernetes.io/projected/c5adfada-3921-46ab-8be3-b9e7e812d276-kube-api-access-wm5n6\") pod \"csi-hostpathplugin-7pmwt\" (UID: \"c5adfada-3921-46ab-8be3-b9e7e812d276\") " pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.618375 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9728r\" (UniqueName: \"kubernetes.io/projected/13435500-f37a-4b8d-a396-1714f365cbc6-kube-api-access-9728r\") pod \"machine-config-server-bd68z\" (UID: \"13435500-f37a-4b8d-a396-1714f365cbc6\") " pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.637614 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qxz55" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.638485 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.638628 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.138611067 +0000 UTC m=+146.845930084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.638727 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.640006 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.139995695 +0000 UTC m=+146.847314712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: W1207 09:08:49.654004 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod860ad6f7_00fc_4974_a703_389c3ca9d3a0.slice/crio-a5484aa702a402e0ae9e1de50b278d3012001342f450bdde06211c146214702c WatchSource:0}: Error finding container a5484aa702a402e0ae9e1de50b278d3012001342f450bdde06211c146214702c: Status 404 returned error can't find the container with id a5484aa702a402e0ae9e1de50b278d3012001342f450bdde06211c146214702c Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.656803 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.658012 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" event={"ID":"e432ea72-cff7-4826-8001-125443ca158a","Type":"ContainerStarted","Data":"3cbe265b4f04b85a417ffc44dc99e2043453d487d8ae950c9d3505521a5a27a5"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.658050 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" event={"ID":"e432ea72-cff7-4826-8001-125443ca158a","Type":"ContainerStarted","Data":"c82a67a7fb2e239592845179dd455bddd452d20d366586c555edc1143c62a7e4"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.658085 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.658104 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" event={"ID":"e432ea72-cff7-4826-8001-125443ca158a","Type":"ContainerStarted","Data":"b8d85434bb45961936e0c520e88de6295e75c7fd7172e42eac7273582b70a83a"} Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.663490 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bd68z" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.740190 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.741156 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.241142465 +0000 UTC m=+146.948461482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.764081 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.841672 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.842086 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.34207222 +0000 UTC m=+147.049391237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.871093 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.895631 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv"] Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.906534 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qll5w" podStartSLOduration=115.906504892 podStartE2EDuration="1m55.906504892s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:49.90571225 +0000 UTC m=+146.613031267" watchObservedRunningTime="2025-12-07 09:08:49.906504892 +0000 UTC m=+146.613823909" Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.944159 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:49 crc kubenswrapper[4838]: E1207 09:08:49.944737 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.444721241 +0000 UTC m=+147.152040258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:49 crc kubenswrapper[4838]: I1207 09:08:49.949745 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gwtm7"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.032499 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.055623 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.055946 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.555934995 +0000 UTC m=+147.263254012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.155790 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-srd6x"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.155962 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.156192 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.656176991 +0000 UTC m=+147.363496008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.171442 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.171799 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.671787315 +0000 UTC m=+147.379106332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.244581 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7qwhs"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.248204 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.285744 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.286045 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.786031622 +0000 UTC m=+147.493350639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.310443 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-88r9w"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.344323 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" podStartSLOduration=116.344289956 podStartE2EDuration="1m56.344289956s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:50.344089161 +0000 UTC m=+147.051408178" watchObservedRunningTime="2025-12-07 09:08:50.344289956 +0000 UTC m=+147.051608973" Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.348199 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.387046 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.387407 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.887377928 +0000 UTC m=+147.594696945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: W1207 09:08:50.436911 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf760f393_f715_4e80_91ea_b8036cf6a3f8.slice/crio-2f1cb0850d74d02962b7c557c13975abfda0103445feff8bb039825e71e5d330 WatchSource:0}: Error finding container 2f1cb0850d74d02962b7c557c13975abfda0103445feff8bb039825e71e5d330: Status 404 returned error can't find the container with id 2f1cb0850d74d02962b7c557c13975abfda0103445feff8bb039825e71e5d330 Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.492933 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.493250 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:50.993235056 +0000 UTC m=+147.700554073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.533594 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.574038 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xrdpm"] Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.586987 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" podStartSLOduration=117.586973315 podStartE2EDuration="1m57.586973315s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:50.584932179 +0000 UTC m=+147.292251206" watchObservedRunningTime="2025-12-07 09:08:50.586973315 +0000 UTC m=+147.294292332" Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.594631 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.595044 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.095031794 +0000 UTC m=+147.802350821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.652013 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7qwhs" event={"ID":"f760f393-f715-4e80-91ea-b8036cf6a3f8","Type":"ContainerStarted","Data":"2f1cb0850d74d02962b7c557c13975abfda0103445feff8bb039825e71e5d330"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.682735 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bd68z" event={"ID":"13435500-f37a-4b8d-a396-1714f365cbc6","Type":"ContainerStarted","Data":"13928c2c33a237a91da08d6bfe3836eb141338c06e4248dbf52e9f10f4fc04fa"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.688945 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" event={"ID":"29a77e31-52d3-4f97-ad09-4422d5bd523a","Type":"ContainerStarted","Data":"a430085e1a6cbda9e7c237832c0904f4955a327648f1de530fb970ed5d1e7398"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.692283 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" event={"ID":"eb898d9c-aa97-4aa4-a31f-ab8659120558","Type":"ContainerStarted","Data":"7113b36ae6948bdb077da841103d1dd40ee9cf77119fbd39f68199e69134f5f7"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.696711 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.696958 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.196936135 +0000 UTC m=+147.904255152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.697137 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.697619 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.197609263 +0000 UTC m=+147.904928280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.713056 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" event={"ID":"4de2ff5d-8c21-4412-b175-446ec4eec272","Type":"ContainerStarted","Data":"ffcd637ac9e5c89cd5c8bec8206cecf362df4eb7c3a14b1ebdae71650e5837c7"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.719249 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" event={"ID":"95bf0de9-2c4f-4504-a083-49cda7af2c1a","Type":"ContainerStarted","Data":"12ce044fb2bd235a0ab5b6dd46f0ffe7d7ad80cc53f0c3dc44c427e5f5186cc4"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.722160 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-srd6x" event={"ID":"f43b40c1-6db1-4026-a531-a020cb538423","Type":"ContainerStarted","Data":"6b753e09b208e6b71c1278b09bc11b1c07e8179263facee2ca0e8c682454602b"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.724345 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" event={"ID":"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4","Type":"ContainerStarted","Data":"9075f9525d53390c61d7754cf0ca605b0d6ca3a32d079fdb261483d7d1960787"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.726183 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" event={"ID":"df3e0f1c-d750-4b75-8f31-aa4faf37093c","Type":"ContainerStarted","Data":"78e1ae2ef5faccaf228832204c0fbcb02f11f6ac3d7ace8a473d1f9bb1380a52"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.726898 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" event={"ID":"055ff049-72f6-415e-afa6-b262b09be1da","Type":"ContainerStarted","Data":"e83edceee7f8e0898ea68ff9bfa0522aa68a44ba4eacea985f50f3cb7dd59c37"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.738549 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" event={"ID":"15119df3-c69c-4ee2-8a7a-f9373494178d","Type":"ContainerStarted","Data":"ae38ad0eef82e0d210cede53615e04fd5bff27273cae48cdba67bd915b17baec"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.750456 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" event={"ID":"860ad6f7-00fc-4974-a703-389c3ca9d3a0","Type":"ContainerStarted","Data":"a5484aa702a402e0ae9e1de50b278d3012001342f450bdde06211c146214702c"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.783171 4838 generic.go:334] "Generic (PLEG): container finished" podID="c34b4216-fa9a-4fd9-9c00-53e62f4c73a9" containerID="7300bc38f68470dd364a056576c0ac3f2187b06cd74e1008ecb271ed1c76f418" exitCode=0 Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.783300 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" event={"ID":"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9","Type":"ContainerDied","Data":"7300bc38f68470dd364a056576c0ac3f2187b06cd74e1008ecb271ed1c76f418"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.798499 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.799989 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.299972547 +0000 UTC m=+148.007291564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.812607 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" event={"ID":"8cecca3b-390d-4346-9bb4-586c9b7f56ea","Type":"ContainerStarted","Data":"9a2c1ede31da80c4a6784129f12bfaff2c68d0b94a77e363bbb8505a227406f7"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.866638 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" event={"ID":"09c2b041-5634-431c-bc61-5a4418c5296e","Type":"ContainerStarted","Data":"0421b5683d70e425d5343a8f88026444d7fa76a5708ac367c3e495bcdfe37c12"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.867629 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.901631 4838 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-t4fqk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.901680 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.903104 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.904140 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" event={"ID":"1fbf61d9-c3ca-4298-aa49-8f113925aec7","Type":"ContainerStarted","Data":"62709aed84d48df4aa4cb37aab815308eae66e81ee9e2f4f9e1d9fd2998fc67e"} Dec 07 09:08:50 crc kubenswrapper[4838]: E1207 09:08:50.908987 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.408967801 +0000 UTC m=+148.116286818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.924656 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" event={"ID":"21433b7b-de00-4a67-8b98-ac5f7ce89ad8","Type":"ContainerStarted","Data":"0720f1936a54dc2a46dcd8b5b6490f9786e1d4075ae39e977e992eee23632c38"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.924718 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" event={"ID":"21433b7b-de00-4a67-8b98-ac5f7ce89ad8","Type":"ContainerStarted","Data":"4bf780fe52c66d2eda5601cd2d3ab5d38ac3a562efe7ee77c8328c6d6c3c17f6"} Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.925671 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.969154 4838 patch_prober.go:28] interesting pod/console-operator-58897d9998-4zsg8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.969215 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" podUID="21433b7b-de00-4a67-8b98-ac5f7ce89ad8" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Dec 07 09:08:50 crc kubenswrapper[4838]: I1207 09:08:50.983497 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pvp5p" event={"ID":"846f52e4-6f0a-4969-9b1d-8465d3a2b730","Type":"ContainerStarted","Data":"60475310d5a8bf3a73b78bf73ef997191cb1cd2d320312cfd121592892ce60a8"} Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.004583 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.007651 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" event={"ID":"642783e8-b8d6-4f1f-a313-a771a915ffb6","Type":"ContainerStarted","Data":"da97cf00ce2a85ef7e1541c69082da10efef678b5a2e099dc8c38a48530874a7"} Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.013289 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.513269727 +0000 UTC m=+148.220588744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.100382 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" event={"ID":"5356b0b5-8205-45e8-9ebb-0b2c7be780be","Type":"ContainerStarted","Data":"a475a410f0d7dc531636064a7d34e29b9129c349f1f692d49d9fe9c5af817c17"} Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.120337 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.120874 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.620861412 +0000 UTC m=+148.328180429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.150480 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" event={"ID":"5cdc38b1-0713-4b57-af9e-40071d064bc6","Type":"ContainerStarted","Data":"9423bd67e13fd95092b60828d81ac214f6068243a83c12898d9d24997ffe0efb"} Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.224448 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" podStartSLOduration=118.224431699 podStartE2EDuration="1m58.224431699s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:51.201592778 +0000 UTC m=+147.908911825" watchObservedRunningTime="2025-12-07 09:08:51.224431699 +0000 UTC m=+147.931750716" Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.224569 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.224808 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-78x4g" podStartSLOduration=118.224803719 podStartE2EDuration="1m58.224803719s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:51.223411201 +0000 UTC m=+147.930730218" watchObservedRunningTime="2025-12-07 09:08:51.224803719 +0000 UTC m=+147.932122736" Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.224902 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.724886781 +0000 UTC m=+148.432205798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.240175 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.274084 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.276070 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" podStartSLOduration=118.276059873 podStartE2EDuration="1m58.276059873s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:51.273959835 +0000 UTC m=+147.981278852" watchObservedRunningTime="2025-12-07 09:08:51.276059873 +0000 UTC m=+147.983378890" Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.295379 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" event={"ID":"22165712-08e9-47cc-aaab-e5da9f92a0fb","Type":"ContainerStarted","Data":"df7b6311f624e2cd2e5db416b2da828db56cbbce12d0435d48dddb0c93d0c34f"} Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.320352 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.336182 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.336549 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.836534077 +0000 UTC m=+148.543853094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.351711 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-pvp5p" podStartSLOduration=117.351694039 podStartE2EDuration="1m57.351694039s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:51.350834356 +0000 UTC m=+148.058153383" watchObservedRunningTime="2025-12-07 09:08:51.351694039 +0000 UTC m=+148.059013056" Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.418499 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.437861 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-htg8d"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.440170 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.441964 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:51.941947543 +0000 UTC m=+148.649266560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.545584 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.546144 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.046131666 +0000 UTC m=+148.753450683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.649084 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.649423 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.149406915 +0000 UTC m=+148.856725932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.750229 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.751206 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.251195442 +0000 UTC m=+148.958514459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.757033 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.816946 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.824483 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll54w"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.837790 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-plzvk"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.847660 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.851740 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.852071 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.352055245 +0000 UTC m=+149.059374262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.871151 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.878582 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:51 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:51 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:51 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.878646 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.880650 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w"] Dec 07 09:08:51 crc kubenswrapper[4838]: I1207 09:08:51.953375 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:51 crc kubenswrapper[4838]: E1207 09:08:51.953743 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.453725619 +0000 UTC m=+149.161044636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.048210 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qxz55"] Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.059291 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.060011 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.559995258 +0000 UTC m=+149.267314275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.077181 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc"] Dec 07 09:08:52 crc kubenswrapper[4838]: W1207 09:08:52.089624 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21071ba2_96ab_4eab_b2d6_92264e094685.slice/crio-9f147dbcc374e6936799ecb600e799b2c38240053162723c92780dbd02e93597 WatchSource:0}: Error finding container 9f147dbcc374e6936799ecb600e799b2c38240053162723c92780dbd02e93597: Status 404 returned error can't find the container with id 9f147dbcc374e6936799ecb600e799b2c38240053162723c92780dbd02e93597 Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.160917 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.161355 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.661344864 +0000 UTC m=+149.368663881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.201954 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7pmwt"] Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.261859 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.262210 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.762195546 +0000 UTC m=+149.469514563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: W1207 09:08:52.341941 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94a2d3db_cde7_4bad_be14_fe5bb0d3e450.slice/crio-4b05247c481a279acd7079cd8b7cbf6e7ed84f5f643004157664b9126c70c810 WatchSource:0}: Error finding container 4b05247c481a279acd7079cd8b7cbf6e7ed84f5f643004157664b9126c70c810: Status 404 returned error can't find the container with id 4b05247c481a279acd7079cd8b7cbf6e7ed84f5f643004157664b9126c70c810 Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.363100 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.363383 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.863369078 +0000 UTC m=+149.570688095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.365494 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" event={"ID":"4de2ff5d-8c21-4412-b175-446ec4eec272","Type":"ContainerStarted","Data":"bdce022a05542cc35aa50caee5e97c7568e1640a01a8af8b8aa3b5ea4895895d"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.381529 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" event={"ID":"2ec53082-404b-4679-89cf-84130cb32f60","Type":"ContainerStarted","Data":"e3706cfdfce16189e8107a72918f62567fba02ac1164ef2081a1d725b5325685"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.384985 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" event={"ID":"ca3a5d32-c068-4069-8f41-3bad3c040ac0","Type":"ContainerStarted","Data":"6937b68f19a4a60e9088ca6dd780892920997a907a81a0d59174525fb8d62dc2"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.401869 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" event={"ID":"21071ba2-96ab-4eab-b2d6-92264e094685","Type":"ContainerStarted","Data":"9f147dbcc374e6936799ecb600e799b2c38240053162723c92780dbd02e93597"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.407885 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" event={"ID":"eb898d9c-aa97-4aa4-a31f-ab8659120558","Type":"ContainerStarted","Data":"f8d1a6562cb6c5a70ada1dcba6abd6d0a3ceed3c563b943c2350acb67f1676bb"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.419025 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dckpn" podStartSLOduration=119.419013161 podStartE2EDuration="1m59.419013161s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.418195098 +0000 UTC m=+149.125514115" watchObservedRunningTime="2025-12-07 09:08:52.419013161 +0000 UTC m=+149.126332168" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.440189 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" event={"ID":"9038da7b-0b0e-44f6-abda-827d7de86d0d","Type":"ContainerStarted","Data":"cf09ba017909341d0ad19ff5728f084f266a49a2afee6095ec2b2f39ecc6e4e4"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.467191 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.468074 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:52.968047304 +0000 UTC m=+149.675366321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.480098 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bd68z" event={"ID":"13435500-f37a-4b8d-a396-1714f365cbc6","Type":"ContainerStarted","Data":"3a074ab9fb9cb870f45ae5adb926d04dfbe2c5e7792cb0aa40de751e3088f861"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.528396 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bd68z" podStartSLOduration=6.528378034 podStartE2EDuration="6.528378034s" podCreationTimestamp="2025-12-07 09:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.511121325 +0000 UTC m=+149.218440332" watchObservedRunningTime="2025-12-07 09:08:52.528378034 +0000 UTC m=+149.235697051" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.552782 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" event={"ID":"1fbf61d9-c3ca-4298-aa49-8f113925aec7","Type":"ContainerStarted","Data":"6825611a6bfff665b32a53cac6db5b36dd6b3cf6f5b1d60a7e19d2eaad2a4bbe"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.564580 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" event={"ID":"7f7cc25b-f9a8-42a5-b114-929a466c16c9","Type":"ContainerStarted","Data":"60632cb2ba36e3e708b9864542660568d8f4f8ec58dd244c6cbfe7bbb141cd4c"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.568934 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.570616 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.070603033 +0000 UTC m=+149.777922050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.579001 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" event={"ID":"860ad6f7-00fc-4974-a703-389c3ca9d3a0","Type":"ContainerStarted","Data":"335af5c76188659a689ec0103bcdab22b292e8e7e4fa53fb5e7eece2b679b9b1"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.580855 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" event={"ID":"3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c","Type":"ContainerStarted","Data":"ad6bd3c35a2ec27500ca346fbf5131403417a2336037b7c6f46730603daa240e"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.581464 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" event={"ID":"22142f3f-8cea-4691-b589-6f4c682740de","Type":"ContainerStarted","Data":"13168ec85b8952f3f90cc024e70c8b8efe1c54850ea29241f875f315a38b66d8"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.582679 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-srd6x" event={"ID":"f43b40c1-6db1-4026-a531-a020cb538423","Type":"ContainerStarted","Data":"cdbe460e6f6c5e863b7fe31f1f70fab9fe630e7618e0a0d47d6532eee008be0d"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.584549 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-srd6x" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.584614 4838 patch_prober.go:28] interesting pod/downloads-7954f5f757-srd6x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.584640 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-srd6x" podUID="f43b40c1-6db1-4026-a531-a020cb538423" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.585758 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7qwhs" event={"ID":"f760f393-f715-4e80-91ea-b8036cf6a3f8","Type":"ContainerStarted","Data":"f568beb33b77d5ad182d30daa90829b11bd99180fe6d4d09d062b6bddc0a8f1f"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.593962 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" event={"ID":"5356b0b5-8205-45e8-9ebb-0b2c7be780be","Type":"ContainerStarted","Data":"5cc7eec093e6d567e0414762bc4eeb0dcb2dad3a7ff74cf2b647a40f33491607"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.606178 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bdbx5" podStartSLOduration=118.606162909 podStartE2EDuration="1m58.606162909s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.604383091 +0000 UTC m=+149.311702108" watchObservedRunningTime="2025-12-07 09:08:52.606162909 +0000 UTC m=+149.313481926" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.617875 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" event={"ID":"11c62d66-9fdc-480e-b259-dc768c6a04e5","Type":"ContainerStarted","Data":"319af27a25c4d4180a7c562ff60466ea49bfcb06080d44068ea9fef1abd36567"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.638269 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" event={"ID":"b6946c48-726b-4b01-a223-2a71d962694b","Type":"ContainerStarted","Data":"d935186a8efd66ee65a3c2505104bab02fee15338e3f61f1f620771fc9cf6b8f"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.638956 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-srd6x" podStartSLOduration=119.638937591 podStartE2EDuration="1m59.638937591s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.638298693 +0000 UTC m=+149.345617710" watchObservedRunningTime="2025-12-07 09:08:52.638937591 +0000 UTC m=+149.346256608" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.650698 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" event={"ID":"642783e8-b8d6-4f1f-a313-a771a915ffb6","Type":"ContainerStarted","Data":"9e7dbf0e894347cfb958976fec933c56ef71db792c4c3986846e6e39ebcdfca2"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.659427 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qxz55" event={"ID":"9bf3b45c-616f-465f-8619-05d4690d6857","Type":"ContainerStarted","Data":"82e65d5f411110e121280cad9821c79a7be1219aeccf8d7ddd6cc9bd7753d837"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.660128 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" event={"ID":"b2067c5a-8073-4b7a-b929-5a4b5ff6846f","Type":"ContainerStarted","Data":"5f188900263ae32f19e87058dc80f722e49ef05422002a3ab41e9bd52a2f4767"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.661272 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" event={"ID":"22165712-08e9-47cc-aaab-e5da9f92a0fb","Type":"ContainerStarted","Data":"c0a795e9e6624f9a68cb8d5cd7d787bc73e70575850b6f994c254a55ccb8363a"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.662365 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" event={"ID":"e32f31a9-a83c-4935-8bb8-71ea757cb10f","Type":"ContainerStarted","Data":"fc3a00baa62ca9d4d323bff6a92d35907fdff74de75ba4942912ddccee69270e"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.663243 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" event={"ID":"5cdc38b1-0713-4b57-af9e-40071d064bc6","Type":"ContainerStarted","Data":"65a9fc8124ff15beb2c99e336041d7bbd4adedcf65c86a6891161647dc0b8858"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.664152 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xrdpm" event={"ID":"66392b25-1220-47fc-96d2-63e43a34ce70","Type":"ContainerStarted","Data":"9737954c9aa09a3f3ed06b38a8ea572fa244875d6f3693e94bdb97ebec3d18f9"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.664772 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" event={"ID":"3854bccd-8788-4cce-8a17-8d06939f172b","Type":"ContainerStarted","Data":"f46b97aba20e3c583512c2d9947eee42e7e28f3fb06e661f42ec989cd975f48b"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.665964 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" event={"ID":"8cecca3b-390d-4346-9bb4-586c9b7f56ea","Type":"ContainerStarted","Data":"119575f11f12bee3883298b24241615a72c1cd677e394873c943a48300d0cc75"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.668696 4838 generic.go:334] "Generic (PLEG): container finished" podID="df3e0f1c-d750-4b75-8f31-aa4faf37093c" containerID="907d2ce1f642e9e5e0115377605517d1dfbbbf08f3e6240ba045174045f3f197" exitCode=0 Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.670461 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" event={"ID":"df3e0f1c-d750-4b75-8f31-aa4faf37093c","Type":"ContainerDied","Data":"907d2ce1f642e9e5e0115377605517d1dfbbbf08f3e6240ba045174045f3f197"} Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.673458 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.673591 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.173576743 +0000 UTC m=+149.880895760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.674074 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.674991 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.174983151 +0000 UTC m=+149.882302168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.682451 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cmsvz" podStartSLOduration=119.682434784 podStartE2EDuration="1m59.682434784s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.679176135 +0000 UTC m=+149.386495152" watchObservedRunningTime="2025-12-07 09:08:52.682434784 +0000 UTC m=+149.389753801" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.778397 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.779502 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.279488963 +0000 UTC m=+149.986807980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.822687 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pcrsf" podStartSLOduration=118.822666517 podStartE2EDuration="1m58.822666517s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.736369 +0000 UTC m=+149.443688017" watchObservedRunningTime="2025-12-07 09:08:52.822666517 +0000 UTC m=+149.529985534" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.851281 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjwmq" podStartSLOduration=118.851262544 podStartE2EDuration="1m58.851262544s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.825112713 +0000 UTC m=+149.532431740" watchObservedRunningTime="2025-12-07 09:08:52.851262544 +0000 UTC m=+149.558581561" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.877108 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:52 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:52 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:52 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.877191 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.880642 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.881198 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.381184988 +0000 UTC m=+150.088504005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.981364 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.981540 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.481515786 +0000 UTC m=+150.188834803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:52 crc kubenswrapper[4838]: I1207 09:08:52.981847 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:52 crc kubenswrapper[4838]: E1207 09:08:52.982159 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.482145243 +0000 UTC m=+150.189464260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.082582 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.082952 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.582937594 +0000 UTC m=+150.290256611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.184417 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.184809 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.684796864 +0000 UTC m=+150.392115881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.285695 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.286032 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.786017856 +0000 UTC m=+150.493336873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.387518 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.387842 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.887827595 +0000 UTC m=+150.595146612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.488889 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.489068 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.989037467 +0000 UTC m=+150.696356484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.489347 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.489695 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:53.989678844 +0000 UTC m=+150.696997871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.592287 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.593070 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.093046925 +0000 UTC m=+150.800365942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.672774 4838 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-t4fqk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.672865 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.672789 4838 patch_prober.go:28] interesting pod/console-operator-58897d9998-4zsg8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.672927 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" podUID="21433b7b-de00-4a67-8b98-ac5f7ce89ad8" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.686117 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-gwtm7" podStartSLOduration=119.686100915 podStartE2EDuration="1m59.686100915s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:52.858650315 +0000 UTC m=+149.565969332" watchObservedRunningTime="2025-12-07 09:08:53.686100915 +0000 UTC m=+150.393419932" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.708479 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.709786 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.209773589 +0000 UTC m=+150.917092606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.772517 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" event={"ID":"c5adfada-3921-46ab-8be3-b9e7e812d276","Type":"ContainerStarted","Data":"a9b46ef5700980f01b92d4f1fc7050d7db6aa5d7a98d21ed5ca3a131fa5c63f4"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.813122 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.813478 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.313463368 +0000 UTC m=+151.020782385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.817960 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" event={"ID":"b6946c48-726b-4b01-a223-2a71d962694b","Type":"ContainerStarted","Data":"f2c0682e2be4355188ec7d3c7e9b5d8fb9d1f41678e6a96927a3abef55948888"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.818553 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.826394 4838 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6rfh7 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.826445 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" podUID="b6946c48-726b-4b01-a223-2a71d962694b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.851123 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" event={"ID":"ca3a5d32-c068-4069-8f41-3bad3c040ac0","Type":"ContainerStarted","Data":"aeb5f047a634825b69d9e3d2322d24d408b3a51299b988a34ededeaca5764ef1"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.874134 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:53 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:53 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:53 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.874195 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.885743 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" event={"ID":"03b837e6-4cd2-43b3-b6b4-a8908c7fbab4","Type":"ContainerStarted","Data":"f7ed558fca2e4997cb5af8c497cb7133b9d88025b1e6ee565d18023664728a09"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.891939 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.894928 4838 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ll54w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.894972 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.915697 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:53 crc kubenswrapper[4838]: E1207 09:08:53.916388 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.416369167 +0000 UTC m=+151.123688224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.925104 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" event={"ID":"055ff049-72f6-415e-afa6-b262b09be1da","Type":"ContainerStarted","Data":"ae2bb0121c130ea95e63b586e22edfdd7eca4d162f85b1c47296cffd726ba016"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.936591 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" event={"ID":"29a77e31-52d3-4f97-ad09-4422d5bd523a","Type":"ContainerStarted","Data":"388ed5e19294d95ca42c49e91bbb6de3d1e732aa49a3010a93cd7941198a6f89"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.946255 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" event={"ID":"642783e8-b8d6-4f1f-a313-a771a915ffb6","Type":"ContainerStarted","Data":"190a8043a38cd7250e21449de35ee08c797b4af64671c3a55f1a9865add73c65"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.954544 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" event={"ID":"2ec53082-404b-4679-89cf-84130cb32f60","Type":"ContainerStarted","Data":"9211d0bb6e7cecdaa7168fa247562374164ffe6826997cfda321833fe8fa6897"} Dec 07 09:08:53 crc kubenswrapper[4838]: I1207 09:08:53.984100 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.016856 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.017948 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.517929238 +0000 UTC m=+151.225248265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.026216 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" event={"ID":"e32f31a9-a83c-4935-8bb8-71ea757cb10f","Type":"ContainerStarted","Data":"45a7b2663eaa93769f35d85dc98be5952fd56be65a81f731d36b6154fdf1e3bf"} Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.026633 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.028437 4838 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-lcg8w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.028493 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" podUID="e32f31a9-a83c-4935-8bb8-71ea757cb10f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.041423 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" event={"ID":"11c62d66-9fdc-480e-b259-dc768c6a04e5","Type":"ContainerStarted","Data":"580f2f9da5cc086f3c110e37b68671eac79b7c85210619f74a10d3060a0f25eb"} Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.065186 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" event={"ID":"94a2d3db-cde7-4bad-be14-fe5bb0d3e450","Type":"ContainerStarted","Data":"4b05247c481a279acd7079cd8b7cbf6e7ed84f5f643004157664b9126c70c810"} Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.099378 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" event={"ID":"95bf0de9-2c4f-4504-a083-49cda7af2c1a","Type":"ContainerStarted","Data":"fec73db0a7a497cf647c57ec159114ff7dd2ffd6a55c5f2692675eacca62da71"} Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.126883 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.128014 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.628003931 +0000 UTC m=+151.335322948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.143455 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" event={"ID":"7f7cc25b-f9a8-42a5-b114-929a466c16c9","Type":"ContainerStarted","Data":"ef387f0167d6b38537766fb695751b2557e215dd91ee1365ad4161a3902ab6db"} Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.144444 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.146306 4838 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-p246s container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.146347 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" podUID="7f7cc25b-f9a8-42a5-b114-929a466c16c9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.157453 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" event={"ID":"9038da7b-0b0e-44f6-abda-827d7de86d0d","Type":"ContainerStarted","Data":"6810218bca27f587827755046314d8d74d6d2d13bb4fe36cfc40df91c6a40c1b"} Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.228355 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.228717 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.728693069 +0000 UTC m=+151.436012086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.228962 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.231449 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.731430264 +0000 UTC m=+151.438749281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.249224 4838 patch_prober.go:28] interesting pod/downloads-7954f5f757-srd6x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.249266 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-srd6x" podUID="f43b40c1-6db1-4026-a531-a020cb538423" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.257073 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.297571 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4zsg8" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.330115 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.332002 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.831985668 +0000 UTC m=+151.539304685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.432376 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.432895 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:54.932883062 +0000 UTC m=+151.640202079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.447376 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k9jbv" podStartSLOduration=120.447361375 podStartE2EDuration="2m0.447361375s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.446011219 +0000 UTC m=+151.153330246" watchObservedRunningTime="2025-12-07 09:08:54.447361375 +0000 UTC m=+151.154680382" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.447572 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" podStartSLOduration=121.447567221 podStartE2EDuration="2m1.447567221s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.405554869 +0000 UTC m=+151.112873886" watchObservedRunningTime="2025-12-07 09:08:54.447567221 +0000 UTC m=+151.154886238" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.493724 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.493789 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.540586 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.541061 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.041038993 +0000 UTC m=+151.748358020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.560192 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-88r9w" podStartSLOduration=120.560175023 podStartE2EDuration="2m0.560175023s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.50123306 +0000 UTC m=+151.208552077" watchObservedRunningTime="2025-12-07 09:08:54.560175023 +0000 UTC m=+151.267494040" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.561357 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vtrtm" podStartSLOduration=120.561348355 podStartE2EDuration="2m0.561348355s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.559064703 +0000 UTC m=+151.266383720" watchObservedRunningTime="2025-12-07 09:08:54.561348355 +0000 UTC m=+151.268667392" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.642245 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.642668 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.142646166 +0000 UTC m=+151.849965263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.686157 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" podStartSLOduration=120.686140788 podStartE2EDuration="2m0.686140788s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.68473224 +0000 UTC m=+151.392051257" watchObservedRunningTime="2025-12-07 09:08:54.686140788 +0000 UTC m=+151.393459805" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.686248 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qxz55" podStartSLOduration=8.686244131 podStartE2EDuration="8.686244131s" podCreationTimestamp="2025-12-07 09:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.654694473 +0000 UTC m=+151.362013490" watchObservedRunningTime="2025-12-07 09:08:54.686244131 +0000 UTC m=+151.393563138" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.729898 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" podStartSLOduration=121.729883048 podStartE2EDuration="2m1.729883048s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.729076376 +0000 UTC m=+151.436395393" watchObservedRunningTime="2025-12-07 09:08:54.729883048 +0000 UTC m=+151.437202065" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.743191 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.743312 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.243286662 +0000 UTC m=+151.950605679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.743457 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.743716 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.243703643 +0000 UTC m=+151.951022660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.832503 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" podStartSLOduration=120.832484988 podStartE2EDuration="2m0.832484988s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.815056784 +0000 UTC m=+151.522375801" watchObservedRunningTime="2025-12-07 09:08:54.832484988 +0000 UTC m=+151.539804005" Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.850445 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.350422035 +0000 UTC m=+152.057741052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.850535 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.851296 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.851703 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.35168798 +0000 UTC m=+152.059006997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.895674 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:54 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:54 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:54 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.895723 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.933007 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" podStartSLOduration=120.93298767 podStartE2EDuration="2m0.93298767s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.884627065 +0000 UTC m=+151.591946092" watchObservedRunningTime="2025-12-07 09:08:54.93298767 +0000 UTC m=+151.640306677" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.934198 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" podStartSLOduration=120.934190153 podStartE2EDuration="2m0.934190153s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:54.931464569 +0000 UTC m=+151.638783586" watchObservedRunningTime="2025-12-07 09:08:54.934190153 +0000 UTC m=+151.641509170" Dec 07 09:08:54 crc kubenswrapper[4838]: I1207 09:08:54.955960 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:54 crc kubenswrapper[4838]: E1207 09:08:54.956279 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.456263693 +0000 UTC m=+152.163582710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.007194 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jdk8t" podStartSLOduration=121.007172238 podStartE2EDuration="2m1.007172238s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.001801392 +0000 UTC m=+151.709120419" watchObservedRunningTime="2025-12-07 09:08:55.007172238 +0000 UTC m=+151.714491245" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.047246 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" podStartSLOduration=121.047229527 podStartE2EDuration="2m1.047229527s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.047085283 +0000 UTC m=+151.754404300" watchObservedRunningTime="2025-12-07 09:08:55.047229527 +0000 UTC m=+151.754548544" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.058485 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.058842 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.558829952 +0000 UTC m=+152.266148969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.135680 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" podStartSLOduration=122.135664592 podStartE2EDuration="2m2.135664592s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.098158622 +0000 UTC m=+151.805477639" watchObservedRunningTime="2025-12-07 09:08:55.135664592 +0000 UTC m=+151.842983599" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.136052 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-xrdpm" podStartSLOduration=122.136048112 podStartE2EDuration="2m2.136048112s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.13525324 +0000 UTC m=+151.842572267" watchObservedRunningTime="2025-12-07 09:08:55.136048112 +0000 UTC m=+151.843367129" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.159711 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.160240 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.660216909 +0000 UTC m=+152.367535926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.229498 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgbcf" podStartSLOduration=122.229482513 podStartE2EDuration="2m2.229482513s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.228327871 +0000 UTC m=+151.935646888" watchObservedRunningTime="2025-12-07 09:08:55.229482513 +0000 UTC m=+151.936801530" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.253379 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tpxz5" event={"ID":"b2067c5a-8073-4b7a-b929-5a4b5ff6846f","Type":"ContainerStarted","Data":"83dfa98dbeb495fa74f2ca7bcd94d4a4502942a95f0665794018967413f34ccd"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.260784 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.261109 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.761094602 +0000 UTC m=+152.468413619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.261097 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" event={"ID":"df3e0f1c-d750-4b75-8f31-aa4faf37093c","Type":"ContainerStarted","Data":"95dae690cea821e1392315aece0838539cd7f2a2b48bfb12762c74494fe682fc"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.263655 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" event={"ID":"c5adfada-3921-46ab-8be3-b9e7e812d276","Type":"ContainerStarted","Data":"9ac07546ccea47e835a09c68feb2803de19bd8dbbdbd4d6c0b0435ebe056f168"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.265733 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" event={"ID":"1fbf61d9-c3ca-4298-aa49-8f113925aec7","Type":"ContainerStarted","Data":"0ddb5c7d705c26a7848d6f1a41899da7b7cedc64a07e3821fa81c58cc4fe1065"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.267971 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" event={"ID":"c34b4216-fa9a-4fd9-9c00-53e62f4c73a9","Type":"ContainerStarted","Data":"cf9180458f6a4d54c1fc4ef59cb584cd8af27885d9fc491665a18cb7e4924be7"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.269639 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" event={"ID":"3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c","Type":"ContainerStarted","Data":"055ba3502038ec5c8f14ff1cc2cd97a85cb702e6c2113f817453f88562f7d729"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.269775 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" event={"ID":"3751e8ca-b7bd-4cca-8ea4-b3f12f1bd21c","Type":"ContainerStarted","Data":"cbfff8a3276822057ac5dcd4203398fb20177ed1e09b88314a41d97a6e9e75af"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.271195 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" event={"ID":"ca3a5d32-c068-4069-8f41-3bad3c040ac0","Type":"ContainerStarted","Data":"7244b51668b0960e666709b2bae253f1697233932dfec216f424872a37e7b96d"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.271341 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.272689 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-c8jw7" event={"ID":"eb898d9c-aa97-4aa4-a31f-ab8659120558","Type":"ContainerStarted","Data":"c06bd4eb24c39ecc8a53a4586decda4adb5d42d6aeede22b0e2a7eb4a3263db2"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.274549 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qxz55" event={"ID":"9bf3b45c-616f-465f-8619-05d4690d6857","Type":"ContainerStarted","Data":"76b0aea0cf9dcbc35f0df367450c08ee8e4b5560a6fd1df95c2323aa7f31ec7e"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.276831 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" event={"ID":"22142f3f-8cea-4691-b589-6f4c682740de","Type":"ContainerStarted","Data":"9d44ae9b85a9a260fd1d4c1ac78552d8648decbde1141944a508780ec047774d"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.276857 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" event={"ID":"22142f3f-8cea-4691-b589-6f4c682740de","Type":"ContainerStarted","Data":"8909a3c4b9a20f01859d4ea2e0f6019059f08dddea66665b1f3315d12f5ed29d"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.278335 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-plzvk" event={"ID":"21071ba2-96ab-4eab-b2d6-92264e094685","Type":"ContainerStarted","Data":"b51f4720bf8f9a8ae6d1c569bba77fc72c84cf902f2d16036e255cd1bfd3afaa"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.280098 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7qwhs" event={"ID":"f760f393-f715-4e80-91ea-b8036cf6a3f8","Type":"ContainerStarted","Data":"cc80e4e168fd9ab4be7cea0eeeeb0d037ac42624c5bfe8eb0f3a9cdb697e2900"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.280285 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-7qwhs" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.282896 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" event={"ID":"94a2d3db-cde7-4bad-be14-fe5bb0d3e450","Type":"ContainerStarted","Data":"4d41f545215070e6277948c27a3f6a0e1e91a0c214f89ee56262811ae5ba3c99"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.282923 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" event={"ID":"94a2d3db-cde7-4bad-be14-fe5bb0d3e450","Type":"ContainerStarted","Data":"01a738a2a0dfb0cac84ecae9fd9bac984d47479e8772d7b207c26a8a592f57da"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.285197 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" event={"ID":"3854bccd-8788-4cce-8a17-8d06939f172b","Type":"ContainerStarted","Data":"948a67c69d5c4326b50d2b256fc103967ff572d7d019d4a2bfa3cdafb23f1a38"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.285734 4838 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ll54w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.285779 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.291124 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" event={"ID":"2ec53082-404b-4679-89cf-84130cb32f60","Type":"ContainerStarted","Data":"29392dff4a61648e9f82d8bdf06526ae8433307e6267d649a7966cf3c5ec980c"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.292760 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xrdpm" event={"ID":"66392b25-1220-47fc-96d2-63e43a34ce70","Type":"ContainerStarted","Data":"3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18"} Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.293997 4838 patch_prober.go:28] interesting pod/downloads-7954f5f757-srd6x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.294051 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-srd6x" podUID="f43b40c1-6db1-4026-a531-a020cb538423" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.299952 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6rfh7" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.317116 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.326730 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" podStartSLOduration=121.326703726 podStartE2EDuration="2m1.326703726s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.267045764 +0000 UTC m=+151.974364781" watchObservedRunningTime="2025-12-07 09:08:55.326703726 +0000 UTC m=+152.034022743" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.327759 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" podStartSLOduration=122.327751855 podStartE2EDuration="2m2.327751855s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.32610535 +0000 UTC m=+152.033424367" watchObservedRunningTime="2025-12-07 09:08:55.327751855 +0000 UTC m=+152.035070872" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.354150 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p246s" podStartSLOduration=121.354133492 podStartE2EDuration="2m1.354133492s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.352849867 +0000 UTC m=+152.060168884" watchObservedRunningTime="2025-12-07 09:08:55.354133492 +0000 UTC m=+152.061452509" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.362041 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.363754 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.863737263 +0000 UTC m=+152.571056280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.393191 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bjfg" podStartSLOduration=121.393175534 podStartE2EDuration="2m1.393175534s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.391636342 +0000 UTC m=+152.098955359" watchObservedRunningTime="2025-12-07 09:08:55.393175534 +0000 UTC m=+152.100494551" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.442470 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zsdrc" podStartSLOduration=121.442452004 podStartE2EDuration="2m1.442452004s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.440427759 +0000 UTC m=+152.147746766" watchObservedRunningTime="2025-12-07 09:08:55.442452004 +0000 UTC m=+152.149771021" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.473823 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.478411 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:55.97839863 +0000 UTC m=+152.685717647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.552417 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" podStartSLOduration=121.552396362 podStartE2EDuration="2m1.552396362s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.550074349 +0000 UTC m=+152.257393376" watchObservedRunningTime="2025-12-07 09:08:55.552396362 +0000 UTC m=+152.259715379" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.574577 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.575321 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.075288805 +0000 UTC m=+152.782607822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.602578 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-htg8d" podStartSLOduration=121.602559906 podStartE2EDuration="2m1.602559906s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.601208109 +0000 UTC m=+152.308527126" watchObservedRunningTime="2025-12-07 09:08:55.602559906 +0000 UTC m=+152.309878933" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.659206 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzx2t" podStartSLOduration=121.659190136 podStartE2EDuration="2m1.659190136s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.658504448 +0000 UTC m=+152.365823465" watchObservedRunningTime="2025-12-07 09:08:55.659190136 +0000 UTC m=+152.366509153" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.678455 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.678524 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.678545 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.678598 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.678630 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.679998 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.179961061 +0000 UTC m=+152.887280078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.680186 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.685123 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.705776 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.706482 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.726430 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-7qwhs" podStartSLOduration=9.726415194 podStartE2EDuration="9.726415194s" podCreationTimestamp="2025-12-07 09:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.723991838 +0000 UTC m=+152.431310855" watchObservedRunningTime="2025-12-07 09:08:55.726415194 +0000 UTC m=+152.433734211" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.747057 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.770351 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6wmc" podStartSLOduration=122.770335188 podStartE2EDuration="2m2.770335188s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:55.769240299 +0000 UTC m=+152.476559326" watchObservedRunningTime="2025-12-07 09:08:55.770335188 +0000 UTC m=+152.477654195" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.779224 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.779512 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.279497878 +0000 UTC m=+152.986816885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.841886 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.856032 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.881551 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.881862 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.381849171 +0000 UTC m=+153.089168188 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.891062 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:55 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:55 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:55 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.891119 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:55 crc kubenswrapper[4838]: I1207 09:08:55.985350 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:55 crc kubenswrapper[4838]: E1207 09:08:55.985738 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.485718865 +0000 UTC m=+153.193037882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.089453 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.089781 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.589768444 +0000 UTC m=+153.297087461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.193320 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.193659 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.693636009 +0000 UTC m=+153.400955026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.256934 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fh5rj"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.257767 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.261619 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.293522 4838 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-lcg8w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.293601 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" podUID="e32f31a9-a83c-4935-8bb8-71ea757cb10f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.294297 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.294605 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.794593144 +0000 UTC m=+153.501912161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.300952 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" event={"ID":"c5adfada-3921-46ab-8be3-b9e7e812d276","Type":"ContainerStarted","Data":"fefa99418650181ecc1e6e711afd66f8567d4475778d6ee9d72b0471dade1f72"} Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.301527 4838 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ll54w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.301558 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.304414 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fh5rj"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.395433 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.395858 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-utilities\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.396029 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-catalog-content\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.396057 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxbj6\" (UniqueName: \"kubernetes.io/projected/26766408-4baa-43e9-9a1d-9b0af9fffe61-kube-api-access-sxbj6\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.396847 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.896809693 +0000 UTC m=+153.604128710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.418477 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.422605 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4z85d"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.423608 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.432310 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.443216 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4z85d"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.497000 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.497042 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-utilities\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.497116 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-catalog-content\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.497139 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxbj6\" (UniqueName: \"kubernetes.io/projected/26766408-4baa-43e9-9a1d-9b0af9fffe61-kube-api-access-sxbj6\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.497396 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:56.997378068 +0000 UTC m=+153.704697085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.497508 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-utilities\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.497893 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-catalog-content\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.538777 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxbj6\" (UniqueName: \"kubernetes.io/projected/26766408-4baa-43e9-9a1d-9b0af9fffe61-kube-api-access-sxbj6\") pod \"certified-operators-fh5rj\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.583297 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.601358 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.601567 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-utilities\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.601611 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-catalog-content\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.601651 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdhsd\" (UniqueName: \"kubernetes.io/projected/e3208dc2-6618-440a-816c-cab7c4fc394c-kube-api-access-cdhsd\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.601765 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.101751776 +0000 UTC m=+153.809070793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.702845 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-catalog-content\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.703247 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdhsd\" (UniqueName: \"kubernetes.io/projected/e3208dc2-6618-440a-816c-cab7c4fc394c-kube-api-access-cdhsd\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.703305 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.703381 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-utilities\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.704185 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-utilities\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.704227 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-catalog-content\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.704543 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.204529571 +0000 UTC m=+153.911848588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.743675 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdhsd\" (UniqueName: \"kubernetes.io/projected/e3208dc2-6618-440a-816c-cab7c4fc394c-kube-api-access-cdhsd\") pod \"community-operators-4z85d\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.757302 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vcz8r"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.773237 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.821301 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.821587 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.321571284 +0000 UTC m=+154.028890301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.844958 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vcz8r"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.875850 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sdc4k" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.878472 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-827bx"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.885020 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:56 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:56 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:56 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.885074 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.900863 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.916262 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-827bx"] Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.930278 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-catalog-content\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.930331 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-utilities\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.930372 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:56 crc kubenswrapper[4838]: I1207 09:08:56.930403 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx6x8\" (UniqueName: \"kubernetes.io/projected/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-kube-api-access-zx6x8\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:56 crc kubenswrapper[4838]: E1207 09:08:56.930684 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.43067266 +0000 UTC m=+154.137991677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.031319 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.031488 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-catalog-content\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.031512 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5svt4\" (UniqueName: \"kubernetes.io/projected/ecb83eed-485c-4624-a038-ff5cec605488-kube-api-access-5svt4\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.031539 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-utilities\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.031623 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-utilities\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.031641 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx6x8\" (UniqueName: \"kubernetes.io/projected/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-kube-api-access-zx6x8\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.031672 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-catalog-content\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.032049 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-catalog-content\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.032379 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-utilities\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.032504 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.532475328 +0000 UTC m=+154.239794345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.049097 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.119105 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx6x8\" (UniqueName: \"kubernetes.io/projected/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-kube-api-access-zx6x8\") pod \"certified-operators-vcz8r\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.132699 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-catalog-content\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.133257 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5svt4\" (UniqueName: \"kubernetes.io/projected/ecb83eed-485c-4624-a038-ff5cec605488-kube-api-access-5svt4\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.133293 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-catalog-content\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.133442 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.133503 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-utilities\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.133747 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.633733662 +0000 UTC m=+154.341052679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.133930 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-utilities\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.176540 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5svt4\" (UniqueName: \"kubernetes.io/projected/ecb83eed-485c-4624-a038-ff5cec605488-kube-api-access-5svt4\") pod \"community-operators-827bx\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.183105 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.232350 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.234842 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.235298 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.735280183 +0000 UTC m=+154.442599200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.335894 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.336305 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.83628529 +0000 UTC m=+154.543604317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.355189 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" event={"ID":"c5adfada-3921-46ab-8be3-b9e7e812d276","Type":"ContainerStarted","Data":"5314ec009c051a97cbc26372642b4f5169e915f7127a60eaa87cbdfa4a1272a2"} Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.372646 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4d06764fcb16f284ec1548dc35ea5e6d0a70380ae213f6091f0ec310fa303fca"} Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.372684 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"43921c9057a2bb6a7abbf988df6bbb881ee744c01e027278d1ba6e96656f69e3"} Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.373775 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:08:57 crc kubenswrapper[4838]: W1207 09:08:57.411505 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-ef0e4cef56fb674cefa48e19e97356a3a32b54c1ce8c3e0795610fd3eae45917 WatchSource:0}: Error finding container ef0e4cef56fb674cefa48e19e97356a3a32b54c1ce8c3e0795610fd3eae45917: Status 404 returned error can't find the container with id ef0e4cef56fb674cefa48e19e97356a3a32b54c1ce8c3e0795610fd3eae45917 Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.437585 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.437962 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:57.937947984 +0000 UTC m=+154.645267001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.544572 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.547010 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.046998489 +0000 UTC m=+154.754317496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.647036 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.647930 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.147901413 +0000 UTC m=+154.855220430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.734971 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fh5rj"] Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.748728 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.749118 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.249103945 +0000 UTC m=+154.956422962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.749322 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4z85d"] Dec 07 09:08:57 crc kubenswrapper[4838]: W1207 09:08:57.777911 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3208dc2_6618_440a_816c_cab7c4fc394c.slice/crio-faceeafa7d1285f7176c1343ca208a8a3b3c46f94e1f54caec1f28dad4e32ef7 WatchSource:0}: Error finding container faceeafa7d1285f7176c1343ca208a8a3b3c46f94e1f54caec1f28dad4e32ef7: Status 404 returned error can't find the container with id faceeafa7d1285f7176c1343ca208a8a3b3c46f94e1f54caec1f28dad4e32ef7 Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.850007 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.850514 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.350498962 +0000 UTC m=+155.057817979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.877023 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:57 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:57 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:57 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.877070 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.921128 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.921173 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.940726 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:57 crc kubenswrapper[4838]: I1207 09:08:57.951891 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:57 crc kubenswrapper[4838]: E1207 09:08:57.952258 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.452246259 +0000 UTC m=+155.159565276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.058463 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.059752 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.559737422 +0000 UTC m=+155.267056439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.159845 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.160181 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.660165462 +0000 UTC m=+155.367484479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.260890 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.261067 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.761038045 +0000 UTC m=+155.468357062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.261237 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.261527 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.761512788 +0000 UTC m=+155.468831805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.333651 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-827bx"] Dec 07 09:08:58 crc kubenswrapper[4838]: W1207 09:08:58.346651 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecb83eed_485c_4624_a038_ff5cec605488.slice/crio-8da39b402bfdcd82fccddf7832f0368a434564a8d30fd5d402a6248f308f8dd3 WatchSource:0}: Error finding container 8da39b402bfdcd82fccddf7832f0368a434564a8d30fd5d402a6248f308f8dd3: Status 404 returned error can't find the container with id 8da39b402bfdcd82fccddf7832f0368a434564a8d30fd5d402a6248f308f8dd3 Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.362076 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.362405 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.862376331 +0000 UTC m=+155.569695348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.368684 4838 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.376571 4838 generic.go:334] "Generic (PLEG): container finished" podID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerID="7ad045a7ac4ab70e3a24b2e9b6275ac18c08fa8d3693cdce4771e00f95d54ac7" exitCode=0 Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.376637 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fh5rj" event={"ID":"26766408-4baa-43e9-9a1d-9b0af9fffe61","Type":"ContainerDied","Data":"7ad045a7ac4ab70e3a24b2e9b6275ac18c08fa8d3693cdce4771e00f95d54ac7"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.376666 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fh5rj" event={"ID":"26766408-4baa-43e9-9a1d-9b0af9fffe61","Type":"ContainerStarted","Data":"610d41efb2d6aef6e3b14157aff12e3c060cee9a0e7a5f5cf7133ebab082941e"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.378128 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.380738 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7bf65a403222e35ed1096f95672b4a6e86d8952867d627b5c040fae4bf526008"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.380772 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ef0e4cef56fb674cefa48e19e97356a3a32b54c1ce8c3e0795610fd3eae45917"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.387006 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" event={"ID":"c5adfada-3921-46ab-8be3-b9e7e812d276","Type":"ContainerStarted","Data":"50ed0798c080e3cab9755bd50d1d03da1d5f85edf24a98215cc9ea59fd86add5"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.396336 4838 generic.go:334] "Generic (PLEG): container finished" podID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerID="46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945" exitCode=0 Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.396658 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4z85d" event={"ID":"e3208dc2-6618-440a-816c-cab7c4fc394c","Type":"ContainerDied","Data":"46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.396696 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4z85d" event={"ID":"e3208dc2-6618-440a-816c-cab7c4fc394c","Type":"ContainerStarted","Data":"faceeafa7d1285f7176c1343ca208a8a3b3c46f94e1f54caec1f28dad4e32ef7"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.427110 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e0d024085f840da96b88b47060691a328bae526217c46bfc1669c71dc15b1fd9"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.427153 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5dd53f10babfdecc864bfb78940fccd6c812f8c188d12f75423e285db3798d8d"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.436623 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7kn79"] Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.437642 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.442742 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-827bx" event={"ID":"ecb83eed-485c-4624-a038-ff5cec605488","Type":"ContainerStarted","Data":"8da39b402bfdcd82fccddf7832f0368a434564a8d30fd5d402a6248f308f8dd3"} Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.445398 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.462954 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.463276 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:58.963264604 +0000 UTC m=+155.670583621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.467148 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-ng7p6" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.478619 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kn79"] Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.499887 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vcz8r"] Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.534061 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.534771 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.561265 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.561500 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.563940 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.564208 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-utilities\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.564324 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-catalog-content\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.564398 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ksbk\" (UniqueName: \"kubernetes.io/projected/be195d3f-68b9-47a0-b726-6e7736450daa-kube-api-access-6ksbk\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.565184 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.566887 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.066872141 +0000 UTC m=+155.774191158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.573217 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.573252 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.587652 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.588779 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7pmwt" podStartSLOduration=12.588761917 podStartE2EDuration="12.588761917s" podCreationTimestamp="2025-12-07 09:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:08:58.587892373 +0000 UTC m=+155.295211390" watchObservedRunningTime="2025-12-07 09:08:58.588761917 +0000 UTC m=+155.296080934" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.665293 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed7f9577-eb90-4251-ade8-6d39a65504c5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.665552 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-utilities\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.665571 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f9577-eb90-4251-ade8-6d39a65504c5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.665617 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.665647 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-catalog-content\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.665678 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ksbk\" (UniqueName: \"kubernetes.io/projected/be195d3f-68b9-47a0-b726-6e7736450daa-kube-api-access-6ksbk\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.666133 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.16612349 +0000 UTC m=+155.873442507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.666365 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-utilities\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.666527 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-catalog-content\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.726617 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ksbk\" (UniqueName: \"kubernetes.io/projected/be195d3f-68b9-47a0-b726-6e7736450daa-kube-api-access-6ksbk\") pod \"redhat-marketplace-7kn79\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.763214 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.767317 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.767455 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.267434485 +0000 UTC m=+155.974753502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.767522 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed7f9577-eb90-4251-ade8-6d39a65504c5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.767548 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f9577-eb90-4251-ade8-6d39a65504c5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.767608 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.768619 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f9577-eb90-4251-ade8-6d39a65504c5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.768977 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.268966457 +0000 UTC m=+155.976285474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.805122 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed7f9577-eb90-4251-ade8-6d39a65504c5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.861037 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vcpdh"] Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.879091 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.874426 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.880389 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.874558 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.874491 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.374475636 +0000 UTC m=+156.081794653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.880334 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:58 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:58 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:58 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.880787 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.880992 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.881229 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.381216019 +0000 UTC m=+156.088535036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.936169 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcpdh"] Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.982998 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.983705 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-catalog-content\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.983804 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4fsc\" (UniqueName: \"kubernetes.io/projected/8072346b-d981-44e2-858d-98b54ce21b8e-kube-api-access-k4fsc\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:58 crc kubenswrapper[4838]: I1207 09:08:58.983906 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-utilities\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:58 crc kubenswrapper[4838]: E1207 09:08:58.984634 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.484615331 +0000 UTC m=+156.191934348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.084999 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-catalog-content\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.085045 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4fsc\" (UniqueName: \"kubernetes.io/projected/8072346b-d981-44e2-858d-98b54ce21b8e-kube-api-access-k4fsc\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.085064 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-utilities\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.085100 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:59 crc kubenswrapper[4838]: E1207 09:08:59.085379 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.585367549 +0000 UTC m=+156.292686566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.085808 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-catalog-content\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.086257 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-utilities\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.125204 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4fsc\" (UniqueName: \"kubernetes.io/projected/8072346b-d981-44e2-858d-98b54ce21b8e-kube-api-access-k4fsc\") pod \"redhat-marketplace-vcpdh\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.143142 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.143190 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.145510 4838 patch_prober.go:28] interesting pod/console-f9d7485db-xrdpm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.145504 4838 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-07T09:08:58.368709283Z","Handler":null,"Name":""} Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.145557 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xrdpm" podUID="66392b25-1220-47fc-96d2-63e43a34ce70" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.188448 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:59 crc kubenswrapper[4838]: E1207 09:08:59.188634 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.688604867 +0000 UTC m=+156.395923884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.188769 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:59 crc kubenswrapper[4838]: E1207 09:08:59.190498 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.690483288 +0000 UTC m=+156.397802305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.203256 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lcg8w" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.212677 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.236484 4838 patch_prober.go:28] interesting pod/downloads-7954f5f757-srd6x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.236536 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-srd6x" podUID="f43b40c1-6db1-4026-a531-a020cb538423" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.236925 4838 patch_prober.go:28] interesting pod/downloads-7954f5f757-srd6x container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.236942 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-srd6x" podUID="f43b40c1-6db1-4026-a531-a020cb538423" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.289487 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:59 crc kubenswrapper[4838]: E1207 09:08:59.289939 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.789914031 +0000 UTC m=+156.497233048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.290100 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:59 crc kubenswrapper[4838]: E1207 09:08:59.290459 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 09:08:59.790453136 +0000 UTC m=+156.497772153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-66ndx" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.311544 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.311563 4838 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.311596 4838 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.398323 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.443018 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d58v2"] Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.445113 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.452735 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.456165 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d58v2"] Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.493087 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kn79"] Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.505418 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-utilities\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.505484 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-catalog-content\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.505505 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp9c2\" (UniqueName: \"kubernetes.io/projected/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-kube-api-access-fp9c2\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.507657 4838 generic.go:334] "Generic (PLEG): container finished" podID="ecb83eed-485c-4624-a038-ff5cec605488" containerID="755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1" exitCode=0 Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.507709 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-827bx" event={"ID":"ecb83eed-485c-4624-a038-ff5cec605488","Type":"ContainerDied","Data":"755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1"} Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.511030 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.516257 4838 generic.go:334] "Generic (PLEG): container finished" podID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerID="d7d99d365d79131f6fa182c15a7363e9df0c8cf00166e956284879a4fd688420" exitCode=0 Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.516338 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcz8r" event={"ID":"5bac913b-a6a6-4ef0-babe-9ad5f39ea480","Type":"ContainerDied","Data":"d7d99d365d79131f6fa182c15a7363e9df0c8cf00166e956284879a4fd688420"} Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.516370 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcz8r" event={"ID":"5bac913b-a6a6-4ef0-babe-9ad5f39ea480","Type":"ContainerStarted","Data":"dfeabdc579d1fb39a78fdaf507491466446b9316bdc02053649ef4a9accf7a6c"} Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.522574 4838 generic.go:334] "Generic (PLEG): container finished" podID="11c62d66-9fdc-480e-b259-dc768c6a04e5" containerID="580f2f9da5cc086f3c110e37b68671eac79b7c85210619f74a10d3060a0f25eb" exitCode=0 Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.522825 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" event={"ID":"11c62d66-9fdc-480e-b259-dc768c6a04e5","Type":"ContainerDied","Data":"580f2f9da5cc086f3c110e37b68671eac79b7c85210619f74a10d3060a0f25eb"} Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.532309 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ggzfg" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.606369 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-catalog-content\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.606442 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp9c2\" (UniqueName: \"kubernetes.io/projected/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-kube-api-access-fp9c2\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.606582 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.606637 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-utilities\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.607390 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-catalog-content\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.608133 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-utilities\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.637095 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.637668 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.649052 4838 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.649098 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.669971 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp9c2\" (UniqueName: \"kubernetes.io/projected/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-kube-api-access-fp9c2\") pod \"redhat-operators-d58v2\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.796366 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.830578 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-72w2r"] Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.842311 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.849634 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-72w2r"] Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.900036 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:08:59 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:08:59 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:08:59 crc kubenswrapper[4838]: healthz check failed Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.900378 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.913456 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thtvz\" (UniqueName: \"kubernetes.io/projected/eedc9629-a39f-4e3c-a6b7-8607d29ded18-kube-api-access-thtvz\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.913549 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-utilities\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.913570 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-catalog-content\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:08:59 crc kubenswrapper[4838]: I1207 09:08:59.981981 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-66ndx\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.014839 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thtvz\" (UniqueName: \"kubernetes.io/projected/eedc9629-a39f-4e3c-a6b7-8607d29ded18-kube-api-access-thtvz\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.014896 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-utilities\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.014917 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-catalog-content\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.015330 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-catalog-content\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.016007 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-utilities\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.039219 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.057515 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thtvz\" (UniqueName: \"kubernetes.io/projected/eedc9629-a39f-4e3c-a6b7-8607d29ded18-kube-api-access-thtvz\") pod \"redhat-operators-72w2r\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.090528 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcpdh"] Dec 07 09:09:00 crc kubenswrapper[4838]: W1207 09:09:00.128914 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8072346b_d981_44e2_858d_98b54ce21b8e.slice/crio-1cafefeaceb933f5e9807066f9a737d9ff7e816753e227dccc6e52721dd20b7b WatchSource:0}: Error finding container 1cafefeaceb933f5e9807066f9a737d9ff7e816753e227dccc6e52721dd20b7b: Status 404 returned error can't find the container with id 1cafefeaceb933f5e9807066f9a737d9ff7e816753e227dccc6e52721dd20b7b Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.216111 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.584550 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcpdh" event={"ID":"8072346b-d981-44e2-858d-98b54ce21b8e","Type":"ContainerStarted","Data":"1cafefeaceb933f5e9807066f9a737d9ff7e816753e227dccc6e52721dd20b7b"} Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.595207 4838 generic.go:334] "Generic (PLEG): container finished" podID="be195d3f-68b9-47a0-b726-6e7736450daa" containerID="4a8649caa7c6d90060b543612b71f6d5c4a689a66e6d42fcdd4cebb4b0caef4d" exitCode=0 Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.595919 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kn79" event={"ID":"be195d3f-68b9-47a0-b726-6e7736450daa","Type":"ContainerDied","Data":"4a8649caa7c6d90060b543612b71f6d5c4a689a66e6d42fcdd4cebb4b0caef4d"} Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.595942 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kn79" event={"ID":"be195d3f-68b9-47a0-b726-6e7736450daa","Type":"ContainerStarted","Data":"f87889610e05f6898177930b2b5a50808dde796879f494e4b4f5bbb692541bdc"} Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.632191 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d58v2"] Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.644357 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed7f9577-eb90-4251-ade8-6d39a65504c5","Type":"ContainerStarted","Data":"86aee8488bde8e545c5f914183396a5b184223ea957cf6b61e88d1b925d24eb2"} Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.749939 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-66ndx"] Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.771213 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.772228 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.775272 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.775396 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.788046 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.837338 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.837396 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.872018 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:00 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:00 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:00 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.872077 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.883635 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-72w2r"] Dec 07 09:09:00 crc kubenswrapper[4838]: W1207 09:09:00.935102 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeedc9629_a39f_4e3c_a6b7_8607d29ded18.slice/crio-190416edc49624bbe71349fc24519d0e0d3282e190df9fc0015a3963dbc5911b WatchSource:0}: Error finding container 190416edc49624bbe71349fc24519d0e0d3282e190df9fc0015a3963dbc5911b: Status 404 returned error can't find the container with id 190416edc49624bbe71349fc24519d0e0d3282e190df9fc0015a3963dbc5911b Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.937959 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.938004 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.938198 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:00 crc kubenswrapper[4838]: I1207 09:09:00.975303 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.153170 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.190147 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.241363 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x96fb\" (UniqueName: \"kubernetes.io/projected/11c62d66-9fdc-480e-b259-dc768c6a04e5-kube-api-access-x96fb\") pod \"11c62d66-9fdc-480e-b259-dc768c6a04e5\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.241408 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11c62d66-9fdc-480e-b259-dc768c6a04e5-secret-volume\") pod \"11c62d66-9fdc-480e-b259-dc768c6a04e5\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.241462 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume\") pod \"11c62d66-9fdc-480e-b259-dc768c6a04e5\" (UID: \"11c62d66-9fdc-480e-b259-dc768c6a04e5\") " Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.242800 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume" (OuterVolumeSpecName: "config-volume") pod "11c62d66-9fdc-480e-b259-dc768c6a04e5" (UID: "11c62d66-9fdc-480e-b259-dc768c6a04e5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.252048 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c62d66-9fdc-480e-b259-dc768c6a04e5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "11c62d66-9fdc-480e-b259-dc768c6a04e5" (UID: "11c62d66-9fdc-480e-b259-dc768c6a04e5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.253597 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11c62d66-9fdc-480e-b259-dc768c6a04e5-kube-api-access-x96fb" (OuterVolumeSpecName: "kube-api-access-x96fb") pod "11c62d66-9fdc-480e-b259-dc768c6a04e5" (UID: "11c62d66-9fdc-480e-b259-dc768c6a04e5"). InnerVolumeSpecName "kube-api-access-x96fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.343710 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x96fb\" (UniqueName: \"kubernetes.io/projected/11c62d66-9fdc-480e-b259-dc768c6a04e5-kube-api-access-x96fb\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.344062 4838 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11c62d66-9fdc-480e-b259-dc768c6a04e5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.344075 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11c62d66-9fdc-480e-b259-dc768c6a04e5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.648013 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.720354 4838 generic.go:334] "Generic (PLEG): container finished" podID="ed7f9577-eb90-4251-ade8-6d39a65504c5" containerID="d4aa48ac7df73ded4869b4a703780131093ef0ec8bb8235204286f4d38841195" exitCode=0 Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.720633 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed7f9577-eb90-4251-ade8-6d39a65504c5","Type":"ContainerDied","Data":"d4aa48ac7df73ded4869b4a703780131093ef0ec8bb8235204286f4d38841195"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.732867 4838 generic.go:334] "Generic (PLEG): container finished" podID="8072346b-d981-44e2-858d-98b54ce21b8e" containerID="b124b5bfc44ebbb3d1217e48a0f4e00a0766a8a6eeaafcc662d9042ed8343671" exitCode=0 Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.733587 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcpdh" event={"ID":"8072346b-d981-44e2-858d-98b54ce21b8e","Type":"ContainerDied","Data":"b124b5bfc44ebbb3d1217e48a0f4e00a0766a8a6eeaafcc662d9042ed8343671"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.747447 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" event={"ID":"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb","Type":"ContainerStarted","Data":"7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.747486 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" event={"ID":"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb","Type":"ContainerStarted","Data":"df54c7d011c4f3ad9f55ee68e58eb45339431009519910f3f32d3a14e5a20818"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.748047 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.761240 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" event={"ID":"11c62d66-9fdc-480e-b259-dc768c6a04e5","Type":"ContainerDied","Data":"319af27a25c4d4180a7c562ff60466ea49bfcb06080d44068ea9fef1abd36567"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.761497 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.762278 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="319af27a25c4d4180a7c562ff60466ea49bfcb06080d44068ea9fef1abd36567" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.771113 4838 generic.go:334] "Generic (PLEG): container finished" podID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerID="7b4e529a3c267d796322a4c73d179623f7dc29d4bc327ec4d3ab30d155580f9c" exitCode=0 Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.771184 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d58v2" event={"ID":"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3","Type":"ContainerDied","Data":"7b4e529a3c267d796322a4c73d179623f7dc29d4bc327ec4d3ab30d155580f9c"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.771209 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d58v2" event={"ID":"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3","Type":"ContainerStarted","Data":"922047045a010712980fef48f7b6f2cfcc6b50db30805cdb06e3b52958d2284f"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.775144 4838 generic.go:334] "Generic (PLEG): container finished" podID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerID="83cb956ab050dd0ca73b6dcb2fd1942d32ee651664ac0f11dd796e78aacdf3db" exitCode=0 Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.775194 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72w2r" event={"ID":"eedc9629-a39f-4e3c-a6b7-8607d29ded18","Type":"ContainerDied","Data":"83cb956ab050dd0ca73b6dcb2fd1942d32ee651664ac0f11dd796e78aacdf3db"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.775285 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72w2r" event={"ID":"eedc9629-a39f-4e3c-a6b7-8607d29ded18","Type":"ContainerStarted","Data":"190416edc49624bbe71349fc24519d0e0d3282e190df9fc0015a3963dbc5911b"} Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.811326 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" podStartSLOduration=127.811306663 podStartE2EDuration="2m7.811306663s" podCreationTimestamp="2025-12-07 09:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:09:01.788989046 +0000 UTC m=+158.496308063" watchObservedRunningTime="2025-12-07 09:09:01.811306663 +0000 UTC m=+158.518625680" Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.871135 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:01 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:01 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:01 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:01 crc kubenswrapper[4838]: I1207 09:09:01.871193 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:02 crc kubenswrapper[4838]: I1207 09:09:02.785966 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3e95f637-2543-42d9-8f37-0f5b70c42ec0","Type":"ContainerStarted","Data":"44dbbeb8cb9a2354e2ce19a150cc9e4cb4b5537bbd918c234234c6198b49b0c8"} Dec 07 09:09:02 crc kubenswrapper[4838]: I1207 09:09:02.786289 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3e95f637-2543-42d9-8f37-0f5b70c42ec0","Type":"ContainerStarted","Data":"5dbd19b6f07e8002ea85fc8796580abd6fb290e07fe6f8e3fe67f303c8867b42"} Dec 07 09:09:02 crc kubenswrapper[4838]: I1207 09:09:02.873724 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:02 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:02 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:02 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:02 crc kubenswrapper[4838]: I1207 09:09:02.873774 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.318449 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.506502 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f9577-eb90-4251-ade8-6d39a65504c5-kubelet-dir\") pod \"ed7f9577-eb90-4251-ade8-6d39a65504c5\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.506618 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed7f9577-eb90-4251-ade8-6d39a65504c5-kube-api-access\") pod \"ed7f9577-eb90-4251-ade8-6d39a65504c5\" (UID: \"ed7f9577-eb90-4251-ade8-6d39a65504c5\") " Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.506752 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed7f9577-eb90-4251-ade8-6d39a65504c5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ed7f9577-eb90-4251-ade8-6d39a65504c5" (UID: "ed7f9577-eb90-4251-ade8-6d39a65504c5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.506911 4838 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f9577-eb90-4251-ade8-6d39a65504c5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.548069 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7f9577-eb90-4251-ade8-6d39a65504c5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ed7f9577-eb90-4251-ade8-6d39a65504c5" (UID: "ed7f9577-eb90-4251-ade8-6d39a65504c5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.608544 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed7f9577-eb90-4251-ade8-6d39a65504c5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.814958 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ed7f9577-eb90-4251-ade8-6d39a65504c5","Type":"ContainerDied","Data":"86aee8488bde8e545c5f914183396a5b184223ea957cf6b61e88d1b925d24eb2"} Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.814996 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86aee8488bde8e545c5f914183396a5b184223ea957cf6b61e88d1b925d24eb2" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.815051 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.826957 4838 generic.go:334] "Generic (PLEG): container finished" podID="3e95f637-2543-42d9-8f37-0f5b70c42ec0" containerID="44dbbeb8cb9a2354e2ce19a150cc9e4cb4b5537bbd918c234234c6198b49b0c8" exitCode=0 Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.827701 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3e95f637-2543-42d9-8f37-0f5b70c42ec0","Type":"ContainerDied","Data":"44dbbeb8cb9a2354e2ce19a150cc9e4cb4b5537bbd918c234234c6198b49b0c8"} Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.876883 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:03 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:03 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:03 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:03 crc kubenswrapper[4838]: I1207 09:09:03.876929 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:04 crc kubenswrapper[4838]: I1207 09:09:04.207115 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-7qwhs" Dec 07 09:09:04 crc kubenswrapper[4838]: I1207 09:09:04.872983 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:04 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:04 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:04 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:04 crc kubenswrapper[4838]: I1207 09:09:04.873041 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.346856 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.486235 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kube-api-access\") pod \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.486476 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kubelet-dir\") pod \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\" (UID: \"3e95f637-2543-42d9-8f37-0f5b70c42ec0\") " Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.486861 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3e95f637-2543-42d9-8f37-0f5b70c42ec0" (UID: "3e95f637-2543-42d9-8f37-0f5b70c42ec0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.488407 4838 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.496122 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3e95f637-2543-42d9-8f37-0f5b70c42ec0" (UID: "3e95f637-2543-42d9-8f37-0f5b70c42ec0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.589856 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3e95f637-2543-42d9-8f37-0f5b70c42ec0-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.857801 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3e95f637-2543-42d9-8f37-0f5b70c42ec0","Type":"ContainerDied","Data":"5dbd19b6f07e8002ea85fc8796580abd6fb290e07fe6f8e3fe67f303c8867b42"} Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.857878 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.857871 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dbd19b6f07e8002ea85fc8796580abd6fb290e07fe6f8e3fe67f303c8867b42" Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.869871 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:05 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:05 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:05 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:05 crc kubenswrapper[4838]: I1207 09:09:05.869943 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:06 crc kubenswrapper[4838]: I1207 09:09:06.875424 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:06 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:06 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:06 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:06 crc kubenswrapper[4838]: I1207 09:09:06.875792 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:07 crc kubenswrapper[4838]: I1207 09:09:07.869856 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:07 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:07 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:07 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:07 crc kubenswrapper[4838]: I1207 09:09:07.869933 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:08 crc kubenswrapper[4838]: I1207 09:09:08.872970 4838 patch_prober.go:28] interesting pod/router-default-5444994796-pvp5p container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 09:09:08 crc kubenswrapper[4838]: [-]has-synced failed: reason withheld Dec 07 09:09:08 crc kubenswrapper[4838]: [+]process-running ok Dec 07 09:09:08 crc kubenswrapper[4838]: healthz check failed Dec 07 09:09:08 crc kubenswrapper[4838]: I1207 09:09:08.873013 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pvp5p" podUID="846f52e4-6f0a-4969-9b1d-8465d3a2b730" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 09:09:09 crc kubenswrapper[4838]: I1207 09:09:09.142952 4838 patch_prober.go:28] interesting pod/console-f9d7485db-xrdpm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 07 09:09:09 crc kubenswrapper[4838]: I1207 09:09:09.143011 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xrdpm" podUID="66392b25-1220-47fc-96d2-63e43a34ce70" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 07 09:09:09 crc kubenswrapper[4838]: I1207 09:09:09.237936 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-srd6x" Dec 07 09:09:09 crc kubenswrapper[4838]: I1207 09:09:09.871602 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:09:09 crc kubenswrapper[4838]: I1207 09:09:09.875127 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-pvp5p" Dec 07 09:09:16 crc kubenswrapper[4838]: I1207 09:09:16.341806 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:09:16 crc kubenswrapper[4838]: I1207 09:09:16.363629 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36f02276-0fd8-4d6e-9b23-9aeb7068d8ff-metrics-certs\") pod \"network-metrics-daemon-7nmsx\" (UID: \"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff\") " pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:09:16 crc kubenswrapper[4838]: I1207 09:09:16.435866 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7nmsx" Dec 07 09:09:19 crc kubenswrapper[4838]: I1207 09:09:19.149217 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:09:19 crc kubenswrapper[4838]: I1207 09:09:19.157593 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:09:20 crc kubenswrapper[4838]: I1207 09:09:20.048083 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:09:24 crc kubenswrapper[4838]: I1207 09:09:24.493434 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:09:24 crc kubenswrapper[4838]: I1207 09:09:24.494097 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:09:29 crc kubenswrapper[4838]: I1207 09:09:29.214997 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6v97j" Dec 07 09:09:35 crc kubenswrapper[4838]: I1207 09:09:35.877549 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.537967 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 07 09:09:36 crc kubenswrapper[4838]: E1207 09:09:36.538335 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e95f637-2543-42d9-8f37-0f5b70c42ec0" containerName="pruner" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.538414 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e95f637-2543-42d9-8f37-0f5b70c42ec0" containerName="pruner" Dec 07 09:09:36 crc kubenswrapper[4838]: E1207 09:09:36.538478 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c62d66-9fdc-480e-b259-dc768c6a04e5" containerName="collect-profiles" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.538606 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c62d66-9fdc-480e-b259-dc768c6a04e5" containerName="collect-profiles" Dec 07 09:09:36 crc kubenswrapper[4838]: E1207 09:09:36.538672 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7f9577-eb90-4251-ade8-6d39a65504c5" containerName="pruner" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.538729 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7f9577-eb90-4251-ade8-6d39a65504c5" containerName="pruner" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.538884 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c62d66-9fdc-480e-b259-dc768c6a04e5" containerName="collect-profiles" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.538959 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e95f637-2543-42d9-8f37-0f5b70c42ec0" containerName="pruner" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.539031 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7f9577-eb90-4251-ade8-6d39a65504c5" containerName="pruner" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.539611 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.542749 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.543959 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.555458 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.643616 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25322d5f-09cb-429c-a4c9-b1b3179f6364-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.643683 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25322d5f-09cb-429c-a4c9-b1b3179f6364-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.745297 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25322d5f-09cb-429c-a4c9-b1b3179f6364-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.745521 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25322d5f-09cb-429c-a4c9-b1b3179f6364-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.745794 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25322d5f-09cb-429c-a4c9-b1b3179f6364-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.778935 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25322d5f-09cb-429c-a4c9-b1b3179f6364-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:36 crc kubenswrapper[4838]: I1207 09:09:36.872916 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:38 crc kubenswrapper[4838]: E1207 09:09:38.367297 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 07 09:09:38 crc kubenswrapper[4838]: E1207 09:09:38.367512 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sxbj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fh5rj_openshift-marketplace(26766408-4baa-43e9-9a1d-9b0af9fffe61): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:38 crc kubenswrapper[4838]: E1207 09:09:38.368723 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fh5rj" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" Dec 07 09:09:39 crc kubenswrapper[4838]: E1207 09:09:39.011334 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 07 09:09:39 crc kubenswrapper[4838]: E1207 09:09:39.011501 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4fsc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vcpdh_openshift-marketplace(8072346b-d981-44e2-858d-98b54ce21b8e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:39 crc kubenswrapper[4838]: E1207 09:09:39.012802 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vcpdh" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.356220 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.357024 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.363343 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.516753 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-var-lock\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.516803 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kube-api-access\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.516987 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.618351 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-var-lock\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.618491 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-var-lock\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.618654 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kube-api-access\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.618766 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.618844 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.636520 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kube-api-access\") pod \"installer-9-crc\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:41 crc kubenswrapper[4838]: I1207 09:09:41.685591 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:09:44 crc kubenswrapper[4838]: E1207 09:09:44.374138 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fh5rj" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" Dec 07 09:09:44 crc kubenswrapper[4838]: E1207 09:09:44.374543 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vcpdh" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" Dec 07 09:09:46 crc kubenswrapper[4838]: E1207 09:09:46.186803 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 07 09:09:46 crc kubenswrapper[4838]: E1207 09:09:46.187250 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thtvz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-72w2r_openshift-marketplace(eedc9629-a39f-4e3c-a6b7-8607d29ded18): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:46 crc kubenswrapper[4838]: E1207 09:09:46.188421 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-72w2r" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.375110 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-72w2r" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.480196 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.480514 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cdhsd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4z85d_openshift-marketplace(e3208dc2-6618-440a-816c-cab7c4fc394c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.481676 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4z85d" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.531051 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.531585 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fp9c2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-d58v2_openshift-marketplace(b3ab327e-67a8-4c75-809c-9ecb0d4a9be3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.533730 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-d58v2" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.567148 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.567286 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5svt4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-827bx_openshift-marketplace(ecb83eed-485c-4624-a038-ff5cec605488): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.569126 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-827bx" podUID="ecb83eed-485c-4624-a038-ff5cec605488" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.590265 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.590404 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zx6x8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vcz8r_openshift-marketplace(5bac913b-a6a6-4ef0-babe-9ad5f39ea480): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.591884 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vcz8r" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.601227 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.601348 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ksbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7kn79_openshift-marketplace(be195d3f-68b9-47a0-b726-6e7736450daa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 09:09:47 crc kubenswrapper[4838]: E1207 09:09:47.602626 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7kn79" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" Dec 07 09:09:47 crc kubenswrapper[4838]: I1207 09:09:47.859548 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 07 09:09:47 crc kubenswrapper[4838]: I1207 09:09:47.957439 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7nmsx"] Dec 07 09:09:47 crc kubenswrapper[4838]: I1207 09:09:47.997435 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 07 09:09:48 crc kubenswrapper[4838]: I1207 09:09:48.152078 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" event={"ID":"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff","Type":"ContainerStarted","Data":"2060f3371a8873bd120f5266b56c88132e1f9ae5cea9c09b1164019e16ea525e"} Dec 07 09:09:48 crc kubenswrapper[4838]: I1207 09:09:48.154905 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69","Type":"ContainerStarted","Data":"5f834bd6fee6c7526d6911ccb6c90aee426c30ba36f82fa6b4df7d8279618cdc"} Dec 07 09:09:48 crc kubenswrapper[4838]: I1207 09:09:48.156886 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"25322d5f-09cb-429c-a4c9-b1b3179f6364","Type":"ContainerStarted","Data":"9a5c3aa08d49942072f1a4bedf2fe047138859a71c3d39be2ee282ee2fa0ea6f"} Dec 07 09:09:48 crc kubenswrapper[4838]: E1207 09:09:48.160535 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7kn79" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" Dec 07 09:09:48 crc kubenswrapper[4838]: E1207 09:09:48.160539 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-827bx" podUID="ecb83eed-485c-4624-a038-ff5cec605488" Dec 07 09:09:48 crc kubenswrapper[4838]: E1207 09:09:48.160634 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-d58v2" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" Dec 07 09:09:48 crc kubenswrapper[4838]: E1207 09:09:48.160739 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4z85d" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" Dec 07 09:09:48 crc kubenswrapper[4838]: E1207 09:09:48.160778 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vcz8r" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" Dec 07 09:09:49 crc kubenswrapper[4838]: I1207 09:09:49.163485 4838 generic.go:334] "Generic (PLEG): container finished" podID="25322d5f-09cb-429c-a4c9-b1b3179f6364" containerID="c776a9d937a1d5832e97abee437e7e4800ddbe3228c2e8e85c790426657cf497" exitCode=0 Dec 07 09:09:49 crc kubenswrapper[4838]: I1207 09:09:49.163573 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"25322d5f-09cb-429c-a4c9-b1b3179f6364","Type":"ContainerDied","Data":"c776a9d937a1d5832e97abee437e7e4800ddbe3228c2e8e85c790426657cf497"} Dec 07 09:09:49 crc kubenswrapper[4838]: I1207 09:09:49.167430 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" event={"ID":"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff","Type":"ContainerStarted","Data":"cd7cc5f613e61d3cd81c2b4703912821c329798e17a9fe803f08f88cc2c78b31"} Dec 07 09:09:49 crc kubenswrapper[4838]: I1207 09:09:49.167499 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7nmsx" event={"ID":"36f02276-0fd8-4d6e-9b23-9aeb7068d8ff","Type":"ContainerStarted","Data":"c8e2d1cc80dca04fe8f8156eac6a240c5097ae1d27a333f582f92810fc6273df"} Dec 07 09:09:49 crc kubenswrapper[4838]: I1207 09:09:49.168962 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69","Type":"ContainerStarted","Data":"a2f493a9d434581010623e981cdf931229b2f11610de0a285fbdc4c36cfb4f44"} Dec 07 09:09:49 crc kubenswrapper[4838]: I1207 09:09:49.201939 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.201917936 podStartE2EDuration="8.201917936s" podCreationTimestamp="2025-12-07 09:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:09:49.198119598 +0000 UTC m=+205.905438615" watchObservedRunningTime="2025-12-07 09:09:49.201917936 +0000 UTC m=+205.909236953" Dec 07 09:09:49 crc kubenswrapper[4838]: I1207 09:09:49.215443 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7nmsx" podStartSLOduration=176.215424891 podStartE2EDuration="2m56.215424891s" podCreationTimestamp="2025-12-07 09:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:09:49.212634031 +0000 UTC m=+205.919953048" watchObservedRunningTime="2025-12-07 09:09:49.215424891 +0000 UTC m=+205.922743908" Dec 07 09:09:50 crc kubenswrapper[4838]: I1207 09:09:50.389700 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:50 crc kubenswrapper[4838]: I1207 09:09:50.538922 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25322d5f-09cb-429c-a4c9-b1b3179f6364-kubelet-dir\") pod \"25322d5f-09cb-429c-a4c9-b1b3179f6364\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " Dec 07 09:09:50 crc kubenswrapper[4838]: I1207 09:09:50.539044 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25322d5f-09cb-429c-a4c9-b1b3179f6364-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "25322d5f-09cb-429c-a4c9-b1b3179f6364" (UID: "25322d5f-09cb-429c-a4c9-b1b3179f6364"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:09:50 crc kubenswrapper[4838]: I1207 09:09:50.539081 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25322d5f-09cb-429c-a4c9-b1b3179f6364-kube-api-access\") pod \"25322d5f-09cb-429c-a4c9-b1b3179f6364\" (UID: \"25322d5f-09cb-429c-a4c9-b1b3179f6364\") " Dec 07 09:09:50 crc kubenswrapper[4838]: I1207 09:09:50.539244 4838 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/25322d5f-09cb-429c-a4c9-b1b3179f6364-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:50 crc kubenswrapper[4838]: I1207 09:09:50.547852 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25322d5f-09cb-429c-a4c9-b1b3179f6364-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "25322d5f-09cb-429c-a4c9-b1b3179f6364" (UID: "25322d5f-09cb-429c-a4c9-b1b3179f6364"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:09:50 crc kubenswrapper[4838]: I1207 09:09:50.640513 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25322d5f-09cb-429c-a4c9-b1b3179f6364-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:09:51 crc kubenswrapper[4838]: I1207 09:09:51.182476 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"25322d5f-09cb-429c-a4c9-b1b3179f6364","Type":"ContainerDied","Data":"9a5c3aa08d49942072f1a4bedf2fe047138859a71c3d39be2ee282ee2fa0ea6f"} Dec 07 09:09:51 crc kubenswrapper[4838]: I1207 09:09:51.182959 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a5c3aa08d49942072f1a4bedf2fe047138859a71c3d39be2ee282ee2fa0ea6f" Dec 07 09:09:51 crc kubenswrapper[4838]: I1207 09:09:51.182630 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 09:09:54 crc kubenswrapper[4838]: I1207 09:09:54.492840 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:09:54 crc kubenswrapper[4838]: I1207 09:09:54.493481 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:09:54 crc kubenswrapper[4838]: I1207 09:09:54.493547 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:09:54 crc kubenswrapper[4838]: I1207 09:09:54.494289 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:09:54 crc kubenswrapper[4838]: I1207 09:09:54.498290 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67" gracePeriod=600 Dec 07 09:09:55 crc kubenswrapper[4838]: I1207 09:09:55.204615 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67" exitCode=0 Dec 07 09:09:55 crc kubenswrapper[4838]: I1207 09:09:55.204689 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67"} Dec 07 09:09:55 crc kubenswrapper[4838]: I1207 09:09:55.205383 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"5071e5fc8badded20c88d60e74e442b4c8646f5320a60585369874896a169785"} Dec 07 09:09:57 crc kubenswrapper[4838]: I1207 09:09:57.228697 4838 generic.go:334] "Generic (PLEG): container finished" podID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerID="39f5cf6816c26f9f4f37667af93943767937025255bc60e93401a6835748aaa8" exitCode=0 Dec 07 09:09:57 crc kubenswrapper[4838]: I1207 09:09:57.229336 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fh5rj" event={"ID":"26766408-4baa-43e9-9a1d-9b0af9fffe61","Type":"ContainerDied","Data":"39f5cf6816c26f9f4f37667af93943767937025255bc60e93401a6835748aaa8"} Dec 07 09:09:58 crc kubenswrapper[4838]: I1207 09:09:58.235650 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fh5rj" event={"ID":"26766408-4baa-43e9-9a1d-9b0af9fffe61","Type":"ContainerStarted","Data":"4f6b7bb02c589b7341245c2556d6dd34da38cb6169f3849ede8b4f89f0aff990"} Dec 07 09:09:58 crc kubenswrapper[4838]: I1207 09:09:58.238154 4838 generic.go:334] "Generic (PLEG): container finished" podID="8072346b-d981-44e2-858d-98b54ce21b8e" containerID="af001461615b25e58e85ebaf9d98f3b11fa5f0524da413a3d5e1e203570afe5a" exitCode=0 Dec 07 09:09:58 crc kubenswrapper[4838]: I1207 09:09:58.238178 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcpdh" event={"ID":"8072346b-d981-44e2-858d-98b54ce21b8e","Type":"ContainerDied","Data":"af001461615b25e58e85ebaf9d98f3b11fa5f0524da413a3d5e1e203570afe5a"} Dec 07 09:09:58 crc kubenswrapper[4838]: I1207 09:09:58.254310 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fh5rj" podStartSLOduration=3.00771901 podStartE2EDuration="1m2.254295443s" podCreationTimestamp="2025-12-07 09:08:56 +0000 UTC" firstStartedPulling="2025-12-07 09:08:58.377894023 +0000 UTC m=+155.085213040" lastFinishedPulling="2025-12-07 09:09:57.624470456 +0000 UTC m=+214.331789473" observedRunningTime="2025-12-07 09:09:58.253052687 +0000 UTC m=+214.960371704" watchObservedRunningTime="2025-12-07 09:09:58.254295443 +0000 UTC m=+214.961614460" Dec 07 09:09:59 crc kubenswrapper[4838]: I1207 09:09:59.246368 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcpdh" event={"ID":"8072346b-d981-44e2-858d-98b54ce21b8e","Type":"ContainerStarted","Data":"336122a48fb2bf6b56ff737e4a457bdf0b088f0770de02a2caad57141974a71e"} Dec 07 09:09:59 crc kubenswrapper[4838]: I1207 09:09:59.270715 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vcpdh" podStartSLOduration=4.406674509 podStartE2EDuration="1m1.270699278s" podCreationTimestamp="2025-12-07 09:08:58 +0000 UTC" firstStartedPulling="2025-12-07 09:09:01.740053375 +0000 UTC m=+158.447372392" lastFinishedPulling="2025-12-07 09:09:58.604078144 +0000 UTC m=+215.311397161" observedRunningTime="2025-12-07 09:09:59.266734295 +0000 UTC m=+215.974053312" watchObservedRunningTime="2025-12-07 09:09:59.270699278 +0000 UTC m=+215.978018295" Dec 07 09:10:00 crc kubenswrapper[4838]: I1207 09:10:00.253359 4838 generic.go:334] "Generic (PLEG): container finished" podID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerID="f9385aca4c27d863ff3e9236bc8e20295eebc1f98c59838e017df3835475a15d" exitCode=0 Dec 07 09:10:00 crc kubenswrapper[4838]: I1207 09:10:00.253428 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcz8r" event={"ID":"5bac913b-a6a6-4ef0-babe-9ad5f39ea480","Type":"ContainerDied","Data":"f9385aca4c27d863ff3e9236bc8e20295eebc1f98c59838e017df3835475a15d"} Dec 07 09:10:00 crc kubenswrapper[4838]: I1207 09:10:00.263869 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72w2r" event={"ID":"eedc9629-a39f-4e3c-a6b7-8607d29ded18","Type":"ContainerStarted","Data":"12687eda214d05e20b4392cb127d8b6a6dfce6cdf4c2fdb9ecdc95f15c16ca02"} Dec 07 09:10:01 crc kubenswrapper[4838]: I1207 09:10:01.271414 4838 generic.go:334] "Generic (PLEG): container finished" podID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerID="12687eda214d05e20b4392cb127d8b6a6dfce6cdf4c2fdb9ecdc95f15c16ca02" exitCode=0 Dec 07 09:10:01 crc kubenswrapper[4838]: I1207 09:10:01.271495 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72w2r" event={"ID":"eedc9629-a39f-4e3c-a6b7-8607d29ded18","Type":"ContainerDied","Data":"12687eda214d05e20b4392cb127d8b6a6dfce6cdf4c2fdb9ecdc95f15c16ca02"} Dec 07 09:10:01 crc kubenswrapper[4838]: I1207 09:10:01.276986 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcz8r" event={"ID":"5bac913b-a6a6-4ef0-babe-9ad5f39ea480","Type":"ContainerStarted","Data":"a3024266b0fdee234f1db516efca55f606db554d2b1b2575ca47abc7c24ac9c9"} Dec 07 09:10:01 crc kubenswrapper[4838]: I1207 09:10:01.307942 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vcz8r" podStartSLOduration=4.196754528 podStartE2EDuration="1m5.307923784s" podCreationTimestamp="2025-12-07 09:08:56 +0000 UTC" firstStartedPulling="2025-12-07 09:08:59.521834688 +0000 UTC m=+156.229153705" lastFinishedPulling="2025-12-07 09:10:00.633003944 +0000 UTC m=+217.340322961" observedRunningTime="2025-12-07 09:10:01.305873516 +0000 UTC m=+218.013192543" watchObservedRunningTime="2025-12-07 09:10:01.307923784 +0000 UTC m=+218.015242811" Dec 07 09:10:02 crc kubenswrapper[4838]: I1207 09:10:02.284058 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72w2r" event={"ID":"eedc9629-a39f-4e3c-a6b7-8607d29ded18","Type":"ContainerStarted","Data":"db6792fe48243ddbd8bcd83e16e23f18ab89aabb197ed293ef25256f65f64956"} Dec 07 09:10:02 crc kubenswrapper[4838]: I1207 09:10:02.301226 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-72w2r" podStartSLOduration=3.385026072 podStartE2EDuration="1m3.301202601s" podCreationTimestamp="2025-12-07 09:08:59 +0000 UTC" firstStartedPulling="2025-12-07 09:09:01.794927997 +0000 UTC m=+158.502247014" lastFinishedPulling="2025-12-07 09:10:01.711104526 +0000 UTC m=+218.418423543" observedRunningTime="2025-12-07 09:10:02.299979787 +0000 UTC m=+219.007298804" watchObservedRunningTime="2025-12-07 09:10:02.301202601 +0000 UTC m=+219.008521628" Dec 07 09:10:04 crc kubenswrapper[4838]: I1207 09:10:04.301698 4838 generic.go:334] "Generic (PLEG): container finished" podID="be195d3f-68b9-47a0-b726-6e7736450daa" containerID="93f4bcabc4a7a75ddc83afe4bfcc46f65205f4610423e76a9141df773f40311c" exitCode=0 Dec 07 09:10:04 crc kubenswrapper[4838]: I1207 09:10:04.302339 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kn79" event={"ID":"be195d3f-68b9-47a0-b726-6e7736450daa","Type":"ContainerDied","Data":"93f4bcabc4a7a75ddc83afe4bfcc46f65205f4610423e76a9141df773f40311c"} Dec 07 09:10:04 crc kubenswrapper[4838]: I1207 09:10:04.307134 4838 generic.go:334] "Generic (PLEG): container finished" podID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerID="a3308a56e96055b755dfa41811e0e3a85d6a494c166703cc0e62158575698d9d" exitCode=0 Dec 07 09:10:04 crc kubenswrapper[4838]: I1207 09:10:04.307220 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d58v2" event={"ID":"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3","Type":"ContainerDied","Data":"a3308a56e96055b755dfa41811e0e3a85d6a494c166703cc0e62158575698d9d"} Dec 07 09:10:04 crc kubenswrapper[4838]: I1207 09:10:04.314583 4838 generic.go:334] "Generic (PLEG): container finished" podID="ecb83eed-485c-4624-a038-ff5cec605488" containerID="88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2" exitCode=0 Dec 07 09:10:04 crc kubenswrapper[4838]: I1207 09:10:04.314616 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-827bx" event={"ID":"ecb83eed-485c-4624-a038-ff5cec605488","Type":"ContainerDied","Data":"88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2"} Dec 07 09:10:05 crc kubenswrapper[4838]: I1207 09:10:05.321162 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4z85d" event={"ID":"e3208dc2-6618-440a-816c-cab7c4fc394c","Type":"ContainerStarted","Data":"47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e"} Dec 07 09:10:06 crc kubenswrapper[4838]: I1207 09:10:06.327405 4838 generic.go:334] "Generic (PLEG): container finished" podID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerID="47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e" exitCode=0 Dec 07 09:10:06 crc kubenswrapper[4838]: I1207 09:10:06.327492 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4z85d" event={"ID":"e3208dc2-6618-440a-816c-cab7c4fc394c","Type":"ContainerDied","Data":"47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e"} Dec 07 09:10:06 crc kubenswrapper[4838]: I1207 09:10:06.584136 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:10:06 crc kubenswrapper[4838]: I1207 09:10:06.584184 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:10:06 crc kubenswrapper[4838]: I1207 09:10:06.692323 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:10:07 crc kubenswrapper[4838]: I1207 09:10:07.183673 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:10:07 crc kubenswrapper[4838]: I1207 09:10:07.184979 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:10:07 crc kubenswrapper[4838]: I1207 09:10:07.217838 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:10:07 crc kubenswrapper[4838]: I1207 09:10:07.372724 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:10:07 crc kubenswrapper[4838]: I1207 09:10:07.379550 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:10:08 crc kubenswrapper[4838]: I1207 09:10:08.231289 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4fqk"] Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.214279 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.214625 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.269067 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.349901 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4z85d" event={"ID":"e3208dc2-6618-440a-816c-cab7c4fc394c","Type":"ContainerStarted","Data":"5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa"} Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.352318 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kn79" event={"ID":"be195d3f-68b9-47a0-b726-6e7736450daa","Type":"ContainerStarted","Data":"9f30edaf5cf0fc6cc6b7c4e4bf076e0e2f44b2d156f4fcb792db41d85e062071"} Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.353920 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d58v2" event={"ID":"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3","Type":"ContainerStarted","Data":"cfe1382645341fe7929cd53ccc106774d41ff28adb85f0b5541b9916a9900dc1"} Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.355718 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-827bx" event={"ID":"ecb83eed-485c-4624-a038-ff5cec605488","Type":"ContainerStarted","Data":"a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70"} Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.399773 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-827bx" podStartSLOduration=4.658754706 podStartE2EDuration="1m13.399754166s" podCreationTimestamp="2025-12-07 09:08:56 +0000 UTC" firstStartedPulling="2025-12-07 09:08:59.521977992 +0000 UTC m=+156.229297009" lastFinishedPulling="2025-12-07 09:10:08.262977462 +0000 UTC m=+224.970296469" observedRunningTime="2025-12-07 09:10:09.398188751 +0000 UTC m=+226.105507768" watchObservedRunningTime="2025-12-07 09:10:09.399754166 +0000 UTC m=+226.107073183" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.400144 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4z85d" podStartSLOduration=3.368314875 podStartE2EDuration="1m13.400139107s" podCreationTimestamp="2025-12-07 09:08:56 +0000 UTC" firstStartedPulling="2025-12-07 09:08:58.404693842 +0000 UTC m=+155.112012859" lastFinishedPulling="2025-12-07 09:10:08.436518074 +0000 UTC m=+225.143837091" observedRunningTime="2025-12-07 09:10:09.381074864 +0000 UTC m=+226.088393871" watchObservedRunningTime="2025-12-07 09:10:09.400139107 +0000 UTC m=+226.107458124" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.417085 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.427061 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d58v2" podStartSLOduration=3.842911224 podStartE2EDuration="1m10.427045933s" podCreationTimestamp="2025-12-07 09:08:59 +0000 UTC" firstStartedPulling="2025-12-07 09:09:01.78618701 +0000 UTC m=+158.493506017" lastFinishedPulling="2025-12-07 09:10:08.370321709 +0000 UTC m=+225.077640726" observedRunningTime="2025-12-07 09:10:09.423973636 +0000 UTC m=+226.131292653" watchObservedRunningTime="2025-12-07 09:10:09.427045933 +0000 UTC m=+226.134364950" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.451606 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7kn79" podStartSLOduration=3.7949543180000003 podStartE2EDuration="1m11.451588422s" podCreationTimestamp="2025-12-07 09:08:58 +0000 UTC" firstStartedPulling="2025-12-07 09:09:00.598455423 +0000 UTC m=+157.305774440" lastFinishedPulling="2025-12-07 09:10:08.255089527 +0000 UTC m=+224.962408544" observedRunningTime="2025-12-07 09:10:09.448472053 +0000 UTC m=+226.155791070" watchObservedRunningTime="2025-12-07 09:10:09.451588422 +0000 UTC m=+226.158907429" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.797096 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:10:09 crc kubenswrapper[4838]: I1207 09:10:09.797147 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:10:10 crc kubenswrapper[4838]: I1207 09:10:10.216943 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:10:10 crc kubenswrapper[4838]: I1207 09:10:10.216997 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:10:10 crc kubenswrapper[4838]: I1207 09:10:10.246367 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vcz8r"] Dec 07 09:10:10 crc kubenswrapper[4838]: I1207 09:10:10.259217 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:10:10 crc kubenswrapper[4838]: I1207 09:10:10.360577 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vcz8r" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="registry-server" containerID="cri-o://a3024266b0fdee234f1db516efca55f606db554d2b1b2575ca47abc7c24ac9c9" gracePeriod=2 Dec 07 09:10:10 crc kubenswrapper[4838]: I1207 09:10:10.417720 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:10:10 crc kubenswrapper[4838]: I1207 09:10:10.836246 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d58v2" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="registry-server" probeResult="failure" output=< Dec 07 09:10:10 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 09:10:10 crc kubenswrapper[4838]: > Dec 07 09:10:11 crc kubenswrapper[4838]: I1207 09:10:11.643762 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcpdh"] Dec 07 09:10:11 crc kubenswrapper[4838]: I1207 09:10:11.644070 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vcpdh" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="registry-server" containerID="cri-o://336122a48fb2bf6b56ff737e4a457bdf0b088f0770de02a2caad57141974a71e" gracePeriod=2 Dec 07 09:10:13 crc kubenswrapper[4838]: I1207 09:10:13.387503 4838 generic.go:334] "Generic (PLEG): container finished" podID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerID="a3024266b0fdee234f1db516efca55f606db554d2b1b2575ca47abc7c24ac9c9" exitCode=0 Dec 07 09:10:13 crc kubenswrapper[4838]: I1207 09:10:13.387643 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcz8r" event={"ID":"5bac913b-a6a6-4ef0-babe-9ad5f39ea480","Type":"ContainerDied","Data":"a3024266b0fdee234f1db516efca55f606db554d2b1b2575ca47abc7c24ac9c9"} Dec 07 09:10:13 crc kubenswrapper[4838]: I1207 09:10:13.391114 4838 generic.go:334] "Generic (PLEG): container finished" podID="8072346b-d981-44e2-858d-98b54ce21b8e" containerID="336122a48fb2bf6b56ff737e4a457bdf0b088f0770de02a2caad57141974a71e" exitCode=0 Dec 07 09:10:13 crc kubenswrapper[4838]: I1207 09:10:13.391175 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcpdh" event={"ID":"8072346b-d981-44e2-858d-98b54ce21b8e","Type":"ContainerDied","Data":"336122a48fb2bf6b56ff737e4a457bdf0b088f0770de02a2caad57141974a71e"} Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.041646 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-72w2r"] Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.044202 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-72w2r" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="registry-server" containerID="cri-o://db6792fe48243ddbd8bcd83e16e23f18ab89aabb197ed293ef25256f65f64956" gracePeriod=2 Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.177332 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.334909 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-utilities\") pod \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.334973 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-catalog-content\") pod \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.335025 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx6x8\" (UniqueName: \"kubernetes.io/projected/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-kube-api-access-zx6x8\") pod \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\" (UID: \"5bac913b-a6a6-4ef0-babe-9ad5f39ea480\") " Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.335750 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-utilities" (OuterVolumeSpecName: "utilities") pod "5bac913b-a6a6-4ef0-babe-9ad5f39ea480" (UID: "5bac913b-a6a6-4ef0-babe-9ad5f39ea480"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.342123 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-kube-api-access-zx6x8" (OuterVolumeSpecName: "kube-api-access-zx6x8") pod "5bac913b-a6a6-4ef0-babe-9ad5f39ea480" (UID: "5bac913b-a6a6-4ef0-babe-9ad5f39ea480"). InnerVolumeSpecName "kube-api-access-zx6x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.361106 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.390834 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bac913b-a6a6-4ef0-babe-9ad5f39ea480" (UID: "5bac913b-a6a6-4ef0-babe-9ad5f39ea480"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.399540 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcpdh" event={"ID":"8072346b-d981-44e2-858d-98b54ce21b8e","Type":"ContainerDied","Data":"1cafefeaceb933f5e9807066f9a737d9ff7e816753e227dccc6e52721dd20b7b"} Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.399623 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcpdh" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.399723 4838 scope.go:117] "RemoveContainer" containerID="336122a48fb2bf6b56ff737e4a457bdf0b088f0770de02a2caad57141974a71e" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.402009 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcz8r" event={"ID":"5bac913b-a6a6-4ef0-babe-9ad5f39ea480","Type":"ContainerDied","Data":"dfeabdc579d1fb39a78fdaf507491466446b9316bdc02053649ef4a9accf7a6c"} Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.402112 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcz8r" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.436561 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-catalog-content\") pod \"8072346b-d981-44e2-858d-98b54ce21b8e\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.436631 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4fsc\" (UniqueName: \"kubernetes.io/projected/8072346b-d981-44e2-858d-98b54ce21b8e-kube-api-access-k4fsc\") pod \"8072346b-d981-44e2-858d-98b54ce21b8e\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.436721 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-utilities\") pod \"8072346b-d981-44e2-858d-98b54ce21b8e\" (UID: \"8072346b-d981-44e2-858d-98b54ce21b8e\") " Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.436972 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.436998 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.437013 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx6x8\" (UniqueName: \"kubernetes.io/projected/5bac913b-a6a6-4ef0-babe-9ad5f39ea480-kube-api-access-zx6x8\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.437610 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-utilities" (OuterVolumeSpecName: "utilities") pod "8072346b-d981-44e2-858d-98b54ce21b8e" (UID: "8072346b-d981-44e2-858d-98b54ce21b8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.439138 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8072346b-d981-44e2-858d-98b54ce21b8e-kube-api-access-k4fsc" (OuterVolumeSpecName: "kube-api-access-k4fsc") pod "8072346b-d981-44e2-858d-98b54ce21b8e" (UID: "8072346b-d981-44e2-858d-98b54ce21b8e"). InnerVolumeSpecName "kube-api-access-k4fsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.454183 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8072346b-d981-44e2-858d-98b54ce21b8e" (UID: "8072346b-d981-44e2-858d-98b54ce21b8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.467059 4838 scope.go:117] "RemoveContainer" containerID="af001461615b25e58e85ebaf9d98f3b11fa5f0524da413a3d5e1e203570afe5a" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.475706 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vcz8r"] Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.478409 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vcz8r"] Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.495602 4838 scope.go:117] "RemoveContainer" containerID="b124b5bfc44ebbb3d1217e48a0f4e00a0766a8a6eeaafcc662d9042ed8343671" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.515663 4838 scope.go:117] "RemoveContainer" containerID="a3024266b0fdee234f1db516efca55f606db554d2b1b2575ca47abc7c24ac9c9" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.531807 4838 scope.go:117] "RemoveContainer" containerID="f9385aca4c27d863ff3e9236bc8e20295eebc1f98c59838e017df3835475a15d" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.538164 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.538187 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4fsc\" (UniqueName: \"kubernetes.io/projected/8072346b-d981-44e2-858d-98b54ce21b8e-kube-api-access-k4fsc\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.538199 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072346b-d981-44e2-858d-98b54ce21b8e-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.545211 4838 scope.go:117] "RemoveContainer" containerID="d7d99d365d79131f6fa182c15a7363e9df0c8cf00166e956284879a4fd688420" Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.733490 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcpdh"] Dec 07 09:10:14 crc kubenswrapper[4838]: I1207 09:10:14.739792 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcpdh"] Dec 07 09:10:15 crc kubenswrapper[4838]: I1207 09:10:15.413311 4838 generic.go:334] "Generic (PLEG): container finished" podID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerID="db6792fe48243ddbd8bcd83e16e23f18ab89aabb197ed293ef25256f65f64956" exitCode=0 Dec 07 09:10:15 crc kubenswrapper[4838]: I1207 09:10:15.413356 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72w2r" event={"ID":"eedc9629-a39f-4e3c-a6b7-8607d29ded18","Type":"ContainerDied","Data":"db6792fe48243ddbd8bcd83e16e23f18ab89aabb197ed293ef25256f65f64956"} Dec 07 09:10:15 crc kubenswrapper[4838]: I1207 09:10:15.622935 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" path="/var/lib/kubelet/pods/5bac913b-a6a6-4ef0-babe-9ad5f39ea480/volumes" Dec 07 09:10:15 crc kubenswrapper[4838]: I1207 09:10:15.624028 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" path="/var/lib/kubelet/pods/8072346b-d981-44e2-858d-98b54ce21b8e/volumes" Dec 07 09:10:15 crc kubenswrapper[4838]: I1207 09:10:15.985794 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.079514 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thtvz\" (UniqueName: \"kubernetes.io/projected/eedc9629-a39f-4e3c-a6b7-8607d29ded18-kube-api-access-thtvz\") pod \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.079554 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-utilities\") pod \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.079662 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-catalog-content\") pod \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\" (UID: \"eedc9629-a39f-4e3c-a6b7-8607d29ded18\") " Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.081001 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-utilities" (OuterVolumeSpecName: "utilities") pod "eedc9629-a39f-4e3c-a6b7-8607d29ded18" (UID: "eedc9629-a39f-4e3c-a6b7-8607d29ded18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.084189 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eedc9629-a39f-4e3c-a6b7-8607d29ded18-kube-api-access-thtvz" (OuterVolumeSpecName: "kube-api-access-thtvz") pod "eedc9629-a39f-4e3c-a6b7-8607d29ded18" (UID: "eedc9629-a39f-4e3c-a6b7-8607d29ded18"). InnerVolumeSpecName "kube-api-access-thtvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.180944 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thtvz\" (UniqueName: \"kubernetes.io/projected/eedc9629-a39f-4e3c-a6b7-8607d29ded18-kube-api-access-thtvz\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.180990 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.183290 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eedc9629-a39f-4e3c-a6b7-8607d29ded18" (UID: "eedc9629-a39f-4e3c-a6b7-8607d29ded18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.282286 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedc9629-a39f-4e3c-a6b7-8607d29ded18-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.426696 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72w2r" event={"ID":"eedc9629-a39f-4e3c-a6b7-8607d29ded18","Type":"ContainerDied","Data":"190416edc49624bbe71349fc24519d0e0d3282e190df9fc0015a3963dbc5911b"} Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.426744 4838 scope.go:117] "RemoveContainer" containerID="db6792fe48243ddbd8bcd83e16e23f18ab89aabb197ed293ef25256f65f64956" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.426803 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72w2r" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.460648 4838 scope.go:117] "RemoveContainer" containerID="12687eda214d05e20b4392cb127d8b6a6dfce6cdf4c2fdb9ecdc95f15c16ca02" Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.477804 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-72w2r"] Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.487331 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-72w2r"] Dec 07 09:10:16 crc kubenswrapper[4838]: I1207 09:10:16.501362 4838 scope.go:117] "RemoveContainer" containerID="83cb956ab050dd0ca73b6dcb2fd1942d32ee651664ac0f11dd796e78aacdf3db" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.050712 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.050770 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.092792 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.232702 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.232808 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.274525 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.474529 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.484057 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:10:17 crc kubenswrapper[4838]: I1207 09:10:17.620404 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" path="/var/lib/kubelet/pods/eedc9629-a39f-4e3c-a6b7-8607d29ded18/volumes" Dec 07 09:10:18 crc kubenswrapper[4838]: I1207 09:10:18.764187 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:10:18 crc kubenswrapper[4838]: I1207 09:10:18.764277 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:10:18 crc kubenswrapper[4838]: I1207 09:10:18.823774 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:10:19 crc kubenswrapper[4838]: I1207 09:10:19.526629 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:10:19 crc kubenswrapper[4838]: I1207 09:10:19.864620 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:10:19 crc kubenswrapper[4838]: I1207 09:10:19.925729 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:10:20 crc kubenswrapper[4838]: I1207 09:10:20.648317 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-827bx"] Dec 07 09:10:20 crc kubenswrapper[4838]: I1207 09:10:20.648660 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-827bx" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="registry-server" containerID="cri-o://a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70" gracePeriod=2 Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.082891 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.281141 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-catalog-content\") pod \"ecb83eed-485c-4624-a038-ff5cec605488\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.281235 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5svt4\" (UniqueName: \"kubernetes.io/projected/ecb83eed-485c-4624-a038-ff5cec605488-kube-api-access-5svt4\") pod \"ecb83eed-485c-4624-a038-ff5cec605488\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.281307 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-utilities\") pod \"ecb83eed-485c-4624-a038-ff5cec605488\" (UID: \"ecb83eed-485c-4624-a038-ff5cec605488\") " Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.283024 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-utilities" (OuterVolumeSpecName: "utilities") pod "ecb83eed-485c-4624-a038-ff5cec605488" (UID: "ecb83eed-485c-4624-a038-ff5cec605488"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.291751 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb83eed-485c-4624-a038-ff5cec605488-kube-api-access-5svt4" (OuterVolumeSpecName: "kube-api-access-5svt4") pod "ecb83eed-485c-4624-a038-ff5cec605488" (UID: "ecb83eed-485c-4624-a038-ff5cec605488"). InnerVolumeSpecName "kube-api-access-5svt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.370928 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecb83eed-485c-4624-a038-ff5cec605488" (UID: "ecb83eed-485c-4624-a038-ff5cec605488"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.382805 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.383094 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5svt4\" (UniqueName: \"kubernetes.io/projected/ecb83eed-485c-4624-a038-ff5cec605488-kube-api-access-5svt4\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.383219 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb83eed-485c-4624-a038-ff5cec605488-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.464109 4838 generic.go:334] "Generic (PLEG): container finished" podID="ecb83eed-485c-4624-a038-ff5cec605488" containerID="a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70" exitCode=0 Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.464176 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-827bx" event={"ID":"ecb83eed-485c-4624-a038-ff5cec605488","Type":"ContainerDied","Data":"a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70"} Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.464217 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-827bx" event={"ID":"ecb83eed-485c-4624-a038-ff5cec605488","Type":"ContainerDied","Data":"8da39b402bfdcd82fccddf7832f0368a434564a8d30fd5d402a6248f308f8dd3"} Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.464247 4838 scope.go:117] "RemoveContainer" containerID="a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.464420 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-827bx" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.496528 4838 scope.go:117] "RemoveContainer" containerID="88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.512681 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-827bx"] Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.515103 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-827bx"] Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.535149 4838 scope.go:117] "RemoveContainer" containerID="755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.557713 4838 scope.go:117] "RemoveContainer" containerID="a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70" Dec 07 09:10:21 crc kubenswrapper[4838]: E1207 09:10:21.558388 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70\": container with ID starting with a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70 not found: ID does not exist" containerID="a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.558419 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70"} err="failed to get container status \"a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70\": rpc error: code = NotFound desc = could not find container \"a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70\": container with ID starting with a87f8df65c85be1be09da273ff56fd8dd65fd3181467aa94d7c0b448d7232d70 not found: ID does not exist" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.558441 4838 scope.go:117] "RemoveContainer" containerID="88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2" Dec 07 09:10:21 crc kubenswrapper[4838]: E1207 09:10:21.558762 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2\": container with ID starting with 88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2 not found: ID does not exist" containerID="88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.558848 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2"} err="failed to get container status \"88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2\": rpc error: code = NotFound desc = could not find container \"88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2\": container with ID starting with 88337db5184daf6a2a15ca02c721cff59bfb1530f2eab10cda337c516b81dbc2 not found: ID does not exist" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.558889 4838 scope.go:117] "RemoveContainer" containerID="755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1" Dec 07 09:10:21 crc kubenswrapper[4838]: E1207 09:10:21.559303 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1\": container with ID starting with 755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1 not found: ID does not exist" containerID="755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.559355 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1"} err="failed to get container status \"755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1\": rpc error: code = NotFound desc = could not find container \"755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1\": container with ID starting with 755da846c6663218af9b6260993b15c0fb3aa7bb16d39239ac4a87f9ad7f66a1 not found: ID does not exist" Dec 07 09:10:21 crc kubenswrapper[4838]: I1207 09:10:21.626905 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb83eed-485c-4624-a038-ff5cec605488" path="/var/lib/kubelet/pods/ecb83eed-485c-4624-a038-ff5cec605488/volumes" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.731111 4838 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732012 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732045 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732071 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732089 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732117 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732133 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732158 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732174 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732204 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732221 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732245 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732262 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732285 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732303 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="extract-content" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732327 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732344 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732369 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732384 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732408 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732425 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732019 4838 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732451 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25322d5f-09cb-429c-a4c9-b1b3179f6364" containerName="pruner" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732577 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="25322d5f-09cb-429c-a4c9-b1b3179f6364" containerName="pruner" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732619 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732648 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.732690 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.732707 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="extract-utilities" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.733066 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb83eed-485c-4624-a038-ff5cec605488" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.733091 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8072346b-d981-44e2-858d-98b54ce21b8e" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.733109 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bac913b-a6a6-4ef0-babe-9ad5f39ea480" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.733124 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="eedc9629-a39f-4e3c-a6b7-8607d29ded18" containerName="registry-server" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.733143 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="25322d5f-09cb-429c-a4c9-b1b3179f6364" containerName="pruner" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.733737 4838 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.733983 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.734287 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee" gracePeriod=15 Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.734349 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883" gracePeriod=15 Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.734389 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4" gracePeriod=15 Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.734318 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9" gracePeriod=15 Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.734431 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834" gracePeriod=15 Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.736476 4838 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.736998 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737065 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.737090 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737138 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.737165 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737229 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.737251 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737263 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.737314 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737328 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.737349 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737360 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737669 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737707 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737721 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737735 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737753 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.737772 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 09:10:25 crc kubenswrapper[4838]: E1207 09:10:25.737989 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.738006 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863176 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863231 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863261 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863294 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863320 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863374 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863394 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.863461 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964691 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964742 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964776 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964811 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964830 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964863 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964889 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964899 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964872 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964940 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964991 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.964965 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.965052 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.965095 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.965152 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:25 crc kubenswrapper[4838]: I1207 09:10:25.965259 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.498659 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.500293 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.501233 4838 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9" exitCode=0 Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.501258 4838 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834" exitCode=0 Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.501266 4838 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883" exitCode=0 Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.501273 4838 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4" exitCode=2 Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.501366 4838 scope.go:117] "RemoveContainer" containerID="27922aeb3fd111b39226413f0080ce4290de914dd24eb1f371840ff5483c8341" Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.503198 4838 generic.go:334] "Generic (PLEG): container finished" podID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" containerID="a2f493a9d434581010623e981cdf931229b2f11610de0a285fbdc4c36cfb4f44" exitCode=0 Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.503250 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69","Type":"ContainerDied","Data":"a2f493a9d434581010623e981cdf931229b2f11610de0a285fbdc4c36cfb4f44"} Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.504616 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:26 crc kubenswrapper[4838]: I1207 09:10:26.505471 4838 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.513724 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.797130 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.798047 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.991384 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-var-lock\") pod \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.991683 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kube-api-access\") pod \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.991719 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-var-lock" (OuterVolumeSpecName: "var-lock") pod "a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" (UID: "a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.991775 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kubelet-dir\") pod \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\" (UID: \"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69\") " Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.992026 4838 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-var-lock\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.992068 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" (UID: "a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:10:27 crc kubenswrapper[4838]: I1207 09:10:27.996429 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" (UID: "a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.093155 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.093180 4838 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.111876 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.112428 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.112881 4838 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.113044 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295158 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295249 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295281 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295347 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295347 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295452 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295527 4838 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295539 4838 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.295547 4838 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.522043 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69","Type":"ContainerDied","Data":"5f834bd6fee6c7526d6911ccb6c90aee426c30ba36f82fa6b4df7d8279618cdc"} Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.522084 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f834bd6fee6c7526d6911ccb6c90aee426c30ba36f82fa6b4df7d8279618cdc" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.522139 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.526893 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.528199 4838 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee" exitCode=0 Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.528271 4838 scope.go:117] "RemoveContainer" containerID="98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.528596 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.541146 4838 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.541756 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.557299 4838 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.557934 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.560151 4838 scope.go:117] "RemoveContainer" containerID="42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.584158 4838 scope.go:117] "RemoveContainer" containerID="fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.601502 4838 scope.go:117] "RemoveContainer" containerID="408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.617785 4838 scope.go:117] "RemoveContainer" containerID="b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.635635 4838 scope.go:117] "RemoveContainer" containerID="beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.662951 4838 scope.go:117] "RemoveContainer" containerID="98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9" Dec 07 09:10:28 crc kubenswrapper[4838]: E1207 09:10:28.663562 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\": container with ID starting with 98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9 not found: ID does not exist" containerID="98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.663670 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9"} err="failed to get container status \"98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\": rpc error: code = NotFound desc = could not find container \"98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9\": container with ID starting with 98ec92153c8ee47541f341dbb0d3b1d2ea12a0ae176ae4f0629e8035ee531dd9 not found: ID does not exist" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.663706 4838 scope.go:117] "RemoveContainer" containerID="42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834" Dec 07 09:10:28 crc kubenswrapper[4838]: E1207 09:10:28.664110 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\": container with ID starting with 42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834 not found: ID does not exist" containerID="42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.664156 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834"} err="failed to get container status \"42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\": rpc error: code = NotFound desc = could not find container \"42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834\": container with ID starting with 42891ba9ee47fe931dbcb5267366ea17e68bb9f0f875c2cf83f444e0f8443834 not found: ID does not exist" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.664185 4838 scope.go:117] "RemoveContainer" containerID="fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883" Dec 07 09:10:28 crc kubenswrapper[4838]: E1207 09:10:28.664509 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\": container with ID starting with fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883 not found: ID does not exist" containerID="fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.664541 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883"} err="failed to get container status \"fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\": rpc error: code = NotFound desc = could not find container \"fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883\": container with ID starting with fbbe10383f830173e7c92871489e862172a8819342ef6246629eacfd1f9d0883 not found: ID does not exist" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.664565 4838 scope.go:117] "RemoveContainer" containerID="408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4" Dec 07 09:10:28 crc kubenswrapper[4838]: E1207 09:10:28.665274 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\": container with ID starting with 408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4 not found: ID does not exist" containerID="408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.665339 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4"} err="failed to get container status \"408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\": rpc error: code = NotFound desc = could not find container \"408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4\": container with ID starting with 408a50a4515378ea12157913f25d406ad7aeb75d93dbac5bdcad353350740fa4 not found: ID does not exist" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.665379 4838 scope.go:117] "RemoveContainer" containerID="b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee" Dec 07 09:10:28 crc kubenswrapper[4838]: E1207 09:10:28.665955 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\": container with ID starting with b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee not found: ID does not exist" containerID="b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.666002 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee"} err="failed to get container status \"b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\": rpc error: code = NotFound desc = could not find container \"b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee\": container with ID starting with b654dabbfccecbb7c914576549f2dedc42911621a52558f8aa8ce1393473bbee not found: ID does not exist" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.666032 4838 scope.go:117] "RemoveContainer" containerID="beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27" Dec 07 09:10:28 crc kubenswrapper[4838]: E1207 09:10:28.666415 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\": container with ID starting with beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27 not found: ID does not exist" containerID="beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27" Dec 07 09:10:28 crc kubenswrapper[4838]: I1207 09:10:28.666452 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27"} err="failed to get container status \"beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\": rpc error: code = NotFound desc = could not find container \"beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27\": container with ID starting with beb4dfc72b7faeb411a7a0fd79024b5775cbd08bfd3c5d6826ebebdbca001e27 not found: ID does not exist" Dec 07 09:10:29 crc kubenswrapper[4838]: I1207 09:10:29.620743 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.295853 4838 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.296392 4838 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.296900 4838 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.297481 4838 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.298035 4838 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:30 crc kubenswrapper[4838]: I1207 09:10:30.298237 4838 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.298690 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="200ms" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.500040 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="400ms" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.790652 4838 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:30 crc kubenswrapper[4838]: I1207 09:10:30.791092 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:30 crc kubenswrapper[4838]: W1207 09:10:30.826809 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-9b2c2b3b79660a335f2a983f7476ca165adcf86fb85926bcc437ad411f1e6686 WatchSource:0}: Error finding container 9b2c2b3b79660a335f2a983f7476ca165adcf86fb85926bcc437ad411f1e6686: Status 404 returned error can't find the container with id 9b2c2b3b79660a335f2a983f7476ca165adcf86fb85926bcc437ad411f1e6686 Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.832129 4838 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.73:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187ee3da543fc4e5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-07 09:10:30.831654117 +0000 UTC m=+247.538973154,LastTimestamp:2025-12-07 09:10:30.831654117 +0000 UTC m=+247.538973154,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 07 09:10:30 crc kubenswrapper[4838]: E1207 09:10:30.901032 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="800ms" Dec 07 09:10:31 crc kubenswrapper[4838]: I1207 09:10:31.545025 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b5af2a42d89c68a8af6a11346591dd2706b88e418c5a506a6da1736aeb682b52"} Dec 07 09:10:31 crc kubenswrapper[4838]: I1207 09:10:31.545352 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"9b2c2b3b79660a335f2a983f7476ca165adcf86fb85926bcc437ad411f1e6686"} Dec 07 09:10:31 crc kubenswrapper[4838]: I1207 09:10:31.546111 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:31 crc kubenswrapper[4838]: E1207 09:10:31.546127 4838 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:10:31 crc kubenswrapper[4838]: E1207 09:10:31.701796 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="1.6s" Dec 07 09:10:32 crc kubenswrapper[4838]: E1207 09:10:32.012631 4838 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.73:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187ee3da543fc4e5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-07 09:10:30.831654117 +0000 UTC m=+247.538973154,LastTimestamp:2025-12-07 09:10:30.831654117 +0000 UTC m=+247.538973154,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.275212 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" containerName="oauth-openshift" containerID="cri-o://0421b5683d70e425d5343a8f88026444d7fa76a5708ac367c3e495bcdfe37c12" gracePeriod=15 Dec 07 09:10:33 crc kubenswrapper[4838]: E1207 09:10:33.302950 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="3.2s" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.559746 4838 generic.go:334] "Generic (PLEG): container finished" podID="09c2b041-5634-431c-bc61-5a4418c5296e" containerID="0421b5683d70e425d5343a8f88026444d7fa76a5708ac367c3e495bcdfe37c12" exitCode=0 Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.559812 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" event={"ID":"09c2b041-5634-431c-bc61-5a4418c5296e","Type":"ContainerDied","Data":"0421b5683d70e425d5343a8f88026444d7fa76a5708ac367c3e495bcdfe37c12"} Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.621241 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.743784 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.744459 4838 status_manager.go:851] "Failed to get status for pod" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4fqk\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.744931 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.872976 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-session\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873045 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r82lc\" (UniqueName: \"kubernetes.io/projected/09c2b041-5634-431c-bc61-5a4418c5296e-kube-api-access-r82lc\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873073 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-cliconfig\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873096 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-audit-policies\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873148 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-ocp-branding-template\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873179 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-error\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873215 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-router-certs\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873242 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-trusted-ca-bundle\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873272 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-provider-selection\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873298 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-login\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873326 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09c2b041-5634-431c-bc61-5a4418c5296e-audit-dir\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873347 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-service-ca\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873385 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-idp-0-file-data\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873408 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-serving-cert\") pod \"09c2b041-5634-431c-bc61-5a4418c5296e\" (UID: \"09c2b041-5634-431c-bc61-5a4418c5296e\") " Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.873859 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09c2b041-5634-431c-bc61-5a4418c5296e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.874886 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.874907 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.874982 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.877088 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.880646 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.882902 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.882945 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09c2b041-5634-431c-bc61-5a4418c5296e-kube-api-access-r82lc" (OuterVolumeSpecName: "kube-api-access-r82lc") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "kube-api-access-r82lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.883225 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.883577 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.886550 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.887012 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.889188 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.894153 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "09c2b041-5634-431c-bc61-5a4418c5296e" (UID: "09c2b041-5634-431c-bc61-5a4418c5296e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974366 4838 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974396 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974410 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974422 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974434 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974445 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974456 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974466 4838 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09c2b041-5634-431c-bc61-5a4418c5296e-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974476 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974486 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974498 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974509 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974519 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r82lc\" (UniqueName: \"kubernetes.io/projected/09c2b041-5634-431c-bc61-5a4418c5296e-kube-api-access-r82lc\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:33 crc kubenswrapper[4838]: I1207 09:10:33.974529 4838 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09c2b041-5634-431c-bc61-5a4418c5296e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 07 09:10:34 crc kubenswrapper[4838]: I1207 09:10:34.569242 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" event={"ID":"09c2b041-5634-431c-bc61-5a4418c5296e","Type":"ContainerDied","Data":"87680cf1c7378825fac8b7a94f57090bf4aeef655ce18f9da77a928e3bbede09"} Dec 07 09:10:34 crc kubenswrapper[4838]: I1207 09:10:34.569329 4838 scope.go:117] "RemoveContainer" containerID="0421b5683d70e425d5343a8f88026444d7fa76a5708ac367c3e495bcdfe37c12" Dec 07 09:10:34 crc kubenswrapper[4838]: I1207 09:10:34.569845 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" Dec 07 09:10:34 crc kubenswrapper[4838]: I1207 09:10:34.571322 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:34 crc kubenswrapper[4838]: I1207 09:10:34.571781 4838 status_manager.go:851] "Failed to get status for pod" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4fqk\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:34 crc kubenswrapper[4838]: I1207 09:10:34.601739 4838 status_manager.go:851] "Failed to get status for pod" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4fqk\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:34 crc kubenswrapper[4838]: I1207 09:10:34.602566 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:36 crc kubenswrapper[4838]: E1207 09:10:36.504159 4838 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.73:6443: connect: connection refused" interval="6.4s" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.609733 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.610074 4838 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="9fbb4f752e3f764b7c5c59d91a7965df1fbea45e17d532e43de493d8c4223233" exitCode=1 Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.610115 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"9fbb4f752e3f764b7c5c59d91a7965df1fbea45e17d532e43de493d8c4223233"} Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.610863 4838 scope.go:117] "RemoveContainer" containerID="9fbb4f752e3f764b7c5c59d91a7965df1fbea45e17d532e43de493d8c4223233" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.611132 4838 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.611701 4838 status_manager.go:851] "Failed to get status for pod" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4fqk\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.612043 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.613612 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.614265 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.615862 4838 status_manager.go:851] "Failed to get status for pod" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4fqk\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.616274 4838 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.638927 4838 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.639007 4838 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:39 crc kubenswrapper[4838]: E1207 09:10:39.640267 4838 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.641193 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:39 crc kubenswrapper[4838]: I1207 09:10:39.722998 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.616571 4838 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d99841c29d375eb37bd76809e75c986c96f8971e04c2ee40d0fa54afdea6d3b7" exitCode=0 Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.616646 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d99841c29d375eb37bd76809e75c986c96f8971e04c2ee40d0fa54afdea6d3b7"} Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.616675 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"08a08b1d2f96245ff4ab65b490dbf9bc8fb160245f61a9b192de6a987385aae9"} Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.616985 4838 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.617004 4838 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.617416 4838 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:40 crc kubenswrapper[4838]: E1207 09:10:40.617416 4838 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.617699 4838 status_manager.go:851] "Failed to get status for pod" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4fqk\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.618018 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.620989 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.621076 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"44f18c93a569afb04df6f57bb72cea7c810cb46cfecfcfb1b17f291d41fc4070"} Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.621641 4838 status_manager.go:851] "Failed to get status for pod" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.621976 4838 status_manager.go:851] "Failed to get status for pod" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" pod="openshift-authentication/oauth-openshift-558db77b4-t4fqk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4fqk\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:40 crc kubenswrapper[4838]: I1207 09:10:40.622205 4838 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.73:6443: connect: connection refused" Dec 07 09:10:41 crc kubenswrapper[4838]: I1207 09:10:41.631946 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3b36fa0558f3408c6b25e683a99eb4d2d15854caae5ce84c7747ebcb6e8328ef"} Dec 07 09:10:41 crc kubenswrapper[4838]: I1207 09:10:41.632210 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4c4372f06e89203c7111add054a979107087bdc90711994d95d35c590c3a274d"} Dec 07 09:10:41 crc kubenswrapper[4838]: I1207 09:10:41.632222 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ed3295a234331288a13115f41108d90546d378636d9489a45b28b5f1ffd19416"} Dec 07 09:10:41 crc kubenswrapper[4838]: I1207 09:10:41.632232 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9dd62e54e24a55208b59aba7db5817337d8c5dcd0f63f1af67cb4a579f743521"} Dec 07 09:10:42 crc kubenswrapper[4838]: I1207 09:10:42.638956 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"919e532f64605a30e4675ba63d93e44935d6479f9a7e7d5b6d7365e8ac7d7919"} Dec 07 09:10:42 crc kubenswrapper[4838]: I1207 09:10:42.639237 4838 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:42 crc kubenswrapper[4838]: I1207 09:10:42.639264 4838 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:42 crc kubenswrapper[4838]: I1207 09:10:42.639247 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:44 crc kubenswrapper[4838]: I1207 09:10:44.642425 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:44 crc kubenswrapper[4838]: I1207 09:10:44.642922 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:44 crc kubenswrapper[4838]: I1207 09:10:44.648284 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:44 crc kubenswrapper[4838]: I1207 09:10:44.785783 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:10:47 crc kubenswrapper[4838]: I1207 09:10:47.648309 4838 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:47 crc kubenswrapper[4838]: I1207 09:10:47.667157 4838 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:47 crc kubenswrapper[4838]: I1207 09:10:47.667188 4838 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:47 crc kubenswrapper[4838]: I1207 09:10:47.671178 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:10:47 crc kubenswrapper[4838]: I1207 09:10:47.726550 4838 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="10ecaf8b-daa4-4cdf-bf0d-e72e0f12d21e" Dec 07 09:10:48 crc kubenswrapper[4838]: I1207 09:10:48.671416 4838 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:48 crc kubenswrapper[4838]: I1207 09:10:48.671448 4838 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:10:48 crc kubenswrapper[4838]: I1207 09:10:48.674185 4838 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="10ecaf8b-daa4-4cdf-bf0d-e72e0f12d21e" Dec 07 09:10:49 crc kubenswrapper[4838]: I1207 09:10:49.723366 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:10:49 crc kubenswrapper[4838]: I1207 09:10:49.728809 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:10:50 crc kubenswrapper[4838]: I1207 09:10:50.690206 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 09:10:56 crc kubenswrapper[4838]: I1207 09:10:56.597168 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 07 09:10:57 crc kubenswrapper[4838]: I1207 09:10:57.334887 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 07 09:10:57 crc kubenswrapper[4838]: I1207 09:10:57.363230 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 07 09:10:57 crc kubenswrapper[4838]: I1207 09:10:57.646875 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 07 09:10:57 crc kubenswrapper[4838]: I1207 09:10:57.970896 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 07 09:10:58 crc kubenswrapper[4838]: I1207 09:10:58.196122 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 07 09:10:58 crc kubenswrapper[4838]: I1207 09:10:58.684574 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.072526 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.104918 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.142371 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.164473 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.297701 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.345325 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.445150 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.598308 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.856374 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.907747 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 07 09:10:59 crc kubenswrapper[4838]: I1207 09:10:59.915453 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.116774 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.258759 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.286653 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.492059 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.492210 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.577262 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.608268 4838 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.847038 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.889836 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.923199 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.931045 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.931853 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.965013 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 07 09:11:00 crc kubenswrapper[4838]: I1207 09:11:00.968471 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.025242 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.143323 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.146609 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.296154 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.397492 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.400452 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.473595 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.558708 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.617194 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.870158 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.906615 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 07 09:11:01 crc kubenswrapper[4838]: I1207 09:11:01.933517 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.027531 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.091637 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.116616 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.268682 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.351696 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.498177 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.528778 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.550511 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.603534 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.613623 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.694527 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.704527 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.788579 4838 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.801745 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.869802 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.956517 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 07 09:11:02 crc kubenswrapper[4838]: I1207 09:11:02.958310 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.009176 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.053956 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.109486 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.159670 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.171076 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.245179 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.320035 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.483187 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.632219 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.734454 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.761458 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 07 09:11:03 crc kubenswrapper[4838]: I1207 09:11:03.956621 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.136718 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.146424 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.176885 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.242862 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.272428 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.304379 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.318257 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.375563 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.494500 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.526248 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.628603 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.628948 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.633240 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.714393 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.774231 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.852911 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 07 09:11:04 crc kubenswrapper[4838]: I1207 09:11:04.999412 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.001729 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.041447 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.198439 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.201009 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.213290 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.247583 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.328803 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.407299 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.436610 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.442767 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.447593 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.514501 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.587627 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.621504 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.675043 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.697511 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.753481 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.767915 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.864730 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.878253 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.890269 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.928684 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 07 09:11:05 crc kubenswrapper[4838]: I1207 09:11:05.965103 4838 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.016248 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.066100 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.100174 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.115173 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.151582 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.161466 4838 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.164892 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.174419 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.207648 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.554195 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.606902 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.652331 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.694661 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.777642 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.802267 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.813664 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.813759 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.836877 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.868094 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.893592 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.951919 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 07 09:11:06 crc kubenswrapper[4838]: I1207 09:11:06.981343 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.012392 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.018466 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.033710 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.164415 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.176171 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.193921 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.314485 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.337370 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.392807 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.438745 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.443075 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.450657 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.616731 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.644934 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.737132 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.737781 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.738736 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.778148 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.811883 4838 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.943889 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.953658 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.969884 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.972038 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 07 09:11:07 crc kubenswrapper[4838]: I1207 09:11:07.983625 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.148013 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.155742 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.331532 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.369096 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.397150 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.532367 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.620060 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.628658 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.635151 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.748247 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.853570 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 07 09:11:08 crc kubenswrapper[4838]: I1207 09:11:08.859320 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.011397 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.061783 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.120153 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.128953 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.156781 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.207160 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.254215 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.255358 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.286857 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.322113 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.508990 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.647771 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.692183 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.795638 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.853015 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 07 09:11:09 crc kubenswrapper[4838]: I1207 09:11:09.924681 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.022304 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.065279 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.087977 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.182621 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.220608 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.327436 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.330648 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.368639 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.380583 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.382377 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.405308 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.442435 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.525275 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.619242 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.684017 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.725251 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.751290 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.781181 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.796474 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.803127 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 07 09:11:10 crc kubenswrapper[4838]: I1207 09:11:10.956635 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.020085 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.093152 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.217629 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.250926 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.302116 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.309718 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.427379 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.680345 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.753752 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.867295 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.952443 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 07 09:11:11 crc kubenswrapper[4838]: I1207 09:11:11.977987 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.022352 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.054952 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.064634 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.165493 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.277291 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.358222 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.767701 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.883838 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.930161 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.972392 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.983211 4838 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.986921 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4fqk","openshift-kube-apiserver/kube-apiserver-crc"] Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.986978 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk"] Dec 07 09:11:12 crc kubenswrapper[4838]: E1207 09:11:12.987152 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" containerName="installer" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.987172 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" containerName="installer" Dec 07 09:11:12 crc kubenswrapper[4838]: E1207 09:11:12.987185 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" containerName="oauth-openshift" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.987191 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" containerName="oauth-openshift" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.987291 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" containerName="oauth-openshift" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.987302 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a81c73fe-8bc9-47ce-a9d4-b3333ecc2f69" containerName="installer" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.987416 4838 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.987439 4838 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f2c8a585-7f49-4169-81b8-34cb0d1a0be9" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.987696 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990063 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4ba01873-c46f-4b10-af04-b9ccfb56b007-audit-dir\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990168 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990209 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-audit-policies\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990267 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990332 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990361 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990390 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82n27\" (UniqueName: \"kubernetes.io/projected/4ba01873-c46f-4b10-af04-b9ccfb56b007-kube-api-access-82n27\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990442 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990470 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990495 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-login\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990530 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-error\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990560 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990579 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.990598 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-session\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993095 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993412 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993432 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993610 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993621 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993876 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993882 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.994017 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.994094 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.993607 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 07 09:11:12 crc kubenswrapper[4838]: I1207 09:11:12.994908 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.000538 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.000801 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.003190 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.009521 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.013386 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.021648 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.034647 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=26.034629385 podStartE2EDuration="26.034629385s" podCreationTimestamp="2025-12-07 09:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:11:13.03092297 +0000 UTC m=+289.738241987" watchObservedRunningTime="2025-12-07 09:11:13.034629385 +0000 UTC m=+289.741948402" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092387 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82n27\" (UniqueName: \"kubernetes.io/projected/4ba01873-c46f-4b10-af04-b9ccfb56b007-kube-api-access-82n27\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092459 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092500 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-login\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092552 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092626 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-error\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092666 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092717 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092749 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-session\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092798 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4ba01873-c46f-4b10-af04-b9ccfb56b007-audit-dir\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092879 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092902 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092947 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-audit-policies\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.092976 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.093013 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.094575 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4ba01873-c46f-4b10-af04-b9ccfb56b007-audit-dir\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.096185 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.096241 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-audit-policies\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.096649 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.097899 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.100150 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.101268 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.112799 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82n27\" (UniqueName: \"kubernetes.io/projected/4ba01873-c46f-4b10-af04-b9ccfb56b007-kube-api-access-82n27\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.118641 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.118971 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.119282 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.119626 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-error\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.122984 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-system-session\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.125335 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4ba01873-c46f-4b10-af04-b9ccfb56b007-v4-0-config-user-template-login\") pod \"oauth-openshift-7cc79f59b7-5lhsk\" (UID: \"4ba01873-c46f-4b10-af04-b9ccfb56b007\") " pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.175834 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.313447 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.507386 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk"] Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.632847 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09c2b041-5634-431c-bc61-5a4418c5296e" path="/var/lib/kubelet/pods/09c2b041-5634-431c-bc61-5a4418c5296e/volumes" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.810290 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" event={"ID":"4ba01873-c46f-4b10-af04-b9ccfb56b007","Type":"ContainerStarted","Data":"93bb06bede17fb9a2532aa8b2b7546cc90120c28c11da254fe6267e6b029d8b0"} Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.810596 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" event={"ID":"4ba01873-c46f-4b10-af04-b9ccfb56b007","Type":"ContainerStarted","Data":"5d11aaaed9552a6d7d2e1277d45c99621399da19c61371217aa582b0a5dbbb66"} Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.810843 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.812452 4838 patch_prober.go:28] interesting pod/oauth-openshift-7cc79f59b7-5lhsk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" start-of-body= Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.812561 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" podUID="4ba01873-c46f-4b10-af04-b9ccfb56b007" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" Dec 07 09:11:13 crc kubenswrapper[4838]: I1207 09:11:13.837220 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 07 09:11:14 crc kubenswrapper[4838]: I1207 09:11:14.071323 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 07 09:11:14 crc kubenswrapper[4838]: I1207 09:11:14.242410 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 07 09:11:14 crc kubenswrapper[4838]: I1207 09:11:14.821281 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" Dec 07 09:11:14 crc kubenswrapper[4838]: I1207 09:11:14.839275 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7cc79f59b7-5lhsk" podStartSLOduration=66.83925988 podStartE2EDuration="1m6.83925988s" podCreationTimestamp="2025-12-07 09:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:11:13.832463824 +0000 UTC m=+290.539782841" watchObservedRunningTime="2025-12-07 09:11:14.83925988 +0000 UTC m=+291.546578897" Dec 07 09:11:14 crc kubenswrapper[4838]: I1207 09:11:14.919740 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 07 09:11:21 crc kubenswrapper[4838]: I1207 09:11:21.462976 4838 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 07 09:11:21 crc kubenswrapper[4838]: I1207 09:11:21.463683 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b5af2a42d89c68a8af6a11346591dd2706b88e418c5a506a6da1736aeb682b52" gracePeriod=5 Dec 07 09:11:23 crc kubenswrapper[4838]: I1207 09:11:23.481364 4838 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 07 09:11:26 crc kubenswrapper[4838]: I1207 09:11:26.885413 4838 generic.go:334] "Generic (PLEG): container finished" podID="3854bccd-8788-4cce-8a17-8d06939f172b" containerID="948a67c69d5c4326b50d2b256fc103967ff572d7d019d4a2bfa3cdafb23f1a38" exitCode=0 Dec 07 09:11:26 crc kubenswrapper[4838]: I1207 09:11:26.885547 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" event={"ID":"3854bccd-8788-4cce-8a17-8d06939f172b","Type":"ContainerDied","Data":"948a67c69d5c4326b50d2b256fc103967ff572d7d019d4a2bfa3cdafb23f1a38"} Dec 07 09:11:26 crc kubenswrapper[4838]: I1207 09:11:26.887338 4838 scope.go:117] "RemoveContainer" containerID="948a67c69d5c4326b50d2b256fc103967ff572d7d019d4a2bfa3cdafb23f1a38" Dec 07 09:11:26 crc kubenswrapper[4838]: I1207 09:11:26.887527 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 07 09:11:26 crc kubenswrapper[4838]: I1207 09:11:26.887564 4838 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b5af2a42d89c68a8af6a11346591dd2706b88e418c5a506a6da1736aeb682b52" exitCode=137 Dec 07 09:11:26 crc kubenswrapper[4838]: E1207 09:11:26.962540 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3854bccd_8788_4cce_8a17_8d06939f172b.slice/crio-conmon-948a67c69d5c4326b50d2b256fc103967ff572d7d019d4a2bfa3cdafb23f1a38.scope\": RecentStats: unable to find data in memory cache]" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.061613 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.062139 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169713 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169780 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169844 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169875 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169905 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169929 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169950 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.169979 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.170044 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.170168 4838 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.170178 4838 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.170186 4838 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.170193 4838 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.179641 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.271492 4838 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.620126 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.893449 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.893553 4838 scope.go:117] "RemoveContainer" containerID="b5af2a42d89c68a8af6a11346591dd2706b88e418c5a506a6da1736aeb682b52" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.893606 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.898032 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" event={"ID":"3854bccd-8788-4cce-8a17-8d06939f172b","Type":"ContainerStarted","Data":"3ec4f3c88edacf90710520d8ed393bdf2fcaa4a691fc0048e77f68fb9a2a37d3"} Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.898421 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:11:27 crc kubenswrapper[4838]: I1207 09:11:27.899794 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.717969 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wmz9f"] Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.718680 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" podUID="950f31f5-e378-4a18-96f7-04e2091e2ca3" containerName="controller-manager" containerID="cri-o://775c31257668362f6bc0212b222993cdc1d81972052e5d369c5269c5df450c75" gracePeriod=30 Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.820151 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf"] Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.820873 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" podUID="9b38416c-b5e5-4195-b950-6e449e035c8b" containerName="route-controller-manager" containerID="cri-o://a4f1cd011880020db373ceb43c25b7c5d1479ad3b2fb01b0f19b321701ff471e" gracePeriod=30 Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.969240 4838 generic.go:334] "Generic (PLEG): container finished" podID="950f31f5-e378-4a18-96f7-04e2091e2ca3" containerID="775c31257668362f6bc0212b222993cdc1d81972052e5d369c5269c5df450c75" exitCode=0 Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.969324 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" event={"ID":"950f31f5-e378-4a18-96f7-04e2091e2ca3","Type":"ContainerDied","Data":"775c31257668362f6bc0212b222993cdc1d81972052e5d369c5269c5df450c75"} Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.971393 4838 generic.go:334] "Generic (PLEG): container finished" podID="9b38416c-b5e5-4195-b950-6e449e035c8b" containerID="a4f1cd011880020db373ceb43c25b7c5d1479ad3b2fb01b0f19b321701ff471e" exitCode=0 Dec 07 09:11:36 crc kubenswrapper[4838]: I1207 09:11:36.971415 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" event={"ID":"9b38416c-b5e5-4195-b950-6e449e035c8b","Type":"ContainerDied","Data":"a4f1cd011880020db373ceb43c25b7c5d1479ad3b2fb01b0f19b321701ff471e"} Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.118830 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.172655 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309224 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-config\") pod \"950f31f5-e378-4a18-96f7-04e2091e2ca3\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309269 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-client-ca\") pod \"950f31f5-e378-4a18-96f7-04e2091e2ca3\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309304 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvxgm\" (UniqueName: \"kubernetes.io/projected/9b38416c-b5e5-4195-b950-6e449e035c8b-kube-api-access-cvxgm\") pod \"9b38416c-b5e5-4195-b950-6e449e035c8b\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309343 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-proxy-ca-bundles\") pod \"950f31f5-e378-4a18-96f7-04e2091e2ca3\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309368 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-config\") pod \"9b38416c-b5e5-4195-b950-6e449e035c8b\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309383 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-888xr\" (UniqueName: \"kubernetes.io/projected/950f31f5-e378-4a18-96f7-04e2091e2ca3-kube-api-access-888xr\") pod \"950f31f5-e378-4a18-96f7-04e2091e2ca3\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309452 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950f31f5-e378-4a18-96f7-04e2091e2ca3-serving-cert\") pod \"950f31f5-e378-4a18-96f7-04e2091e2ca3\" (UID: \"950f31f5-e378-4a18-96f7-04e2091e2ca3\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309469 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b38416c-b5e5-4195-b950-6e449e035c8b-serving-cert\") pod \"9b38416c-b5e5-4195-b950-6e449e035c8b\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.309493 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-client-ca\") pod \"9b38416c-b5e5-4195-b950-6e449e035c8b\" (UID: \"9b38416c-b5e5-4195-b950-6e449e035c8b\") " Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.310439 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-config" (OuterVolumeSpecName: "config") pod "9b38416c-b5e5-4195-b950-6e449e035c8b" (UID: "9b38416c-b5e5-4195-b950-6e449e035c8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.310963 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-client-ca" (OuterVolumeSpecName: "client-ca") pod "950f31f5-e378-4a18-96f7-04e2091e2ca3" (UID: "950f31f5-e378-4a18-96f7-04e2091e2ca3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.311058 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.311119 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "950f31f5-e378-4a18-96f7-04e2091e2ca3" (UID: "950f31f5-e378-4a18-96f7-04e2091e2ca3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.311631 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-client-ca" (OuterVolumeSpecName: "client-ca") pod "9b38416c-b5e5-4195-b950-6e449e035c8b" (UID: "9b38416c-b5e5-4195-b950-6e449e035c8b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.311802 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-config" (OuterVolumeSpecName: "config") pod "950f31f5-e378-4a18-96f7-04e2091e2ca3" (UID: "950f31f5-e378-4a18-96f7-04e2091e2ca3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.315368 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/950f31f5-e378-4a18-96f7-04e2091e2ca3-kube-api-access-888xr" (OuterVolumeSpecName: "kube-api-access-888xr") pod "950f31f5-e378-4a18-96f7-04e2091e2ca3" (UID: "950f31f5-e378-4a18-96f7-04e2091e2ca3"). InnerVolumeSpecName "kube-api-access-888xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.315678 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/950f31f5-e378-4a18-96f7-04e2091e2ca3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "950f31f5-e378-4a18-96f7-04e2091e2ca3" (UID: "950f31f5-e378-4a18-96f7-04e2091e2ca3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.316214 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b38416c-b5e5-4195-b950-6e449e035c8b-kube-api-access-cvxgm" (OuterVolumeSpecName: "kube-api-access-cvxgm") pod "9b38416c-b5e5-4195-b950-6e449e035c8b" (UID: "9b38416c-b5e5-4195-b950-6e449e035c8b"). InnerVolumeSpecName "kube-api-access-cvxgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.316220 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b38416c-b5e5-4195-b950-6e449e035c8b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9b38416c-b5e5-4195-b950-6e449e035c8b" (UID: "9b38416c-b5e5-4195-b950-6e449e035c8b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.411878 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950f31f5-e378-4a18-96f7-04e2091e2ca3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.411937 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b38416c-b5e5-4195-b950-6e449e035c8b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.411956 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b38416c-b5e5-4195-b950-6e449e035c8b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.411975 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.411996 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.412013 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvxgm\" (UniqueName: \"kubernetes.io/projected/9b38416c-b5e5-4195-b950-6e449e035c8b-kube-api-access-cvxgm\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.412032 4838 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950f31f5-e378-4a18-96f7-04e2091e2ca3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.412050 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-888xr\" (UniqueName: \"kubernetes.io/projected/950f31f5-e378-4a18-96f7-04e2091e2ca3-kube-api-access-888xr\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.790320 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-lncht"] Dec 07 09:11:37 crc kubenswrapper[4838]: E1207 09:11:37.790571 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.790588 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 07 09:11:37 crc kubenswrapper[4838]: E1207 09:11:37.790606 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="950f31f5-e378-4a18-96f7-04e2091e2ca3" containerName="controller-manager" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.790615 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="950f31f5-e378-4a18-96f7-04e2091e2ca3" containerName="controller-manager" Dec 07 09:11:37 crc kubenswrapper[4838]: E1207 09:11:37.790624 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b38416c-b5e5-4195-b950-6e449e035c8b" containerName="route-controller-manager" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.790632 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b38416c-b5e5-4195-b950-6e449e035c8b" containerName="route-controller-manager" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.790751 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="950f31f5-e378-4a18-96f7-04e2091e2ca3" containerName="controller-manager" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.790767 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.790782 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b38416c-b5e5-4195-b950-6e449e035c8b" containerName="route-controller-manager" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.791329 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.817604 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc"] Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.818722 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.822584 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-lncht"] Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.835047 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc"] Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919493 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-client-ca\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919551 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-config\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919620 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/154fd6b3-3bd9-467d-9435-4417c448f831-serving-cert\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919655 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-proxy-ca-bundles\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919686 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-client-ca\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919711 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-config\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919758 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc5bl\" (UniqueName: \"kubernetes.io/projected/a1cacabd-b318-4200-b357-181cb9bfdc78-kube-api-access-qc5bl\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919782 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmlcw\" (UniqueName: \"kubernetes.io/projected/154fd6b3-3bd9-467d-9435-4417c448f831-kube-api-access-jmlcw\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.919832 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cacabd-b318-4200-b357-181cb9bfdc78-serving-cert\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.978360 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.978359 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wmz9f" event={"ID":"950f31f5-e378-4a18-96f7-04e2091e2ca3","Type":"ContainerDied","Data":"ba28e2fa50928e7e957e63da3cf9cd16552fb5b96f412d3f7241555ec42d1ad9"} Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.978526 4838 scope.go:117] "RemoveContainer" containerID="775c31257668362f6bc0212b222993cdc1d81972052e5d369c5269c5df450c75" Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.983445 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" event={"ID":"9b38416c-b5e5-4195-b950-6e449e035c8b","Type":"ContainerDied","Data":"cd9ede871ca01861ca4285d84d408fab645da056f33270c89a0ddefba73d7ca1"} Dec 07 09:11:37 crc kubenswrapper[4838]: I1207 09:11:37.983551 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.002723 4838 scope.go:117] "RemoveContainer" containerID="a4f1cd011880020db373ceb43c25b7c5d1479ad3b2fb01b0f19b321701ff471e" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.002854 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wmz9f"] Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.008897 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wmz9f"] Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.019920 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf"] Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.020696 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-client-ca\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.020750 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-config\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.020804 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/154fd6b3-3bd9-467d-9435-4417c448f831-serving-cert\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.020872 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-proxy-ca-bundles\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.020924 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-client-ca\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.020964 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-config\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.021006 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc5bl\" (UniqueName: \"kubernetes.io/projected/a1cacabd-b318-4200-b357-181cb9bfdc78-kube-api-access-qc5bl\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.021037 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ccfmf"] Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.021046 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmlcw\" (UniqueName: \"kubernetes.io/projected/154fd6b3-3bd9-467d-9435-4417c448f831-kube-api-access-jmlcw\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.021088 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cacabd-b318-4200-b357-181cb9bfdc78-serving-cert\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.022516 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-proxy-ca-bundles\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.023553 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-client-ca\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.024871 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-config\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.026022 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-client-ca\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.026101 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-config\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.028583 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/154fd6b3-3bd9-467d-9435-4417c448f831-serving-cert\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.037793 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cacabd-b318-4200-b357-181cb9bfdc78-serving-cert\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.053236 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc5bl\" (UniqueName: \"kubernetes.io/projected/a1cacabd-b318-4200-b357-181cb9bfdc78-kube-api-access-qc5bl\") pod \"controller-manager-6df5f84b58-lncht\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.061918 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmlcw\" (UniqueName: \"kubernetes.io/projected/154fd6b3-3bd9-467d-9435-4417c448f831-kube-api-access-jmlcw\") pod \"route-controller-manager-cf78f955c-wp6qc\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.107788 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.165138 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.340045 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-lncht"] Dec 07 09:11:38 crc kubenswrapper[4838]: W1207 09:11:38.346779 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1cacabd_b318_4200_b357_181cb9bfdc78.slice/crio-6651784c1f972e1750612313ea1ac3f179f2b73412516ec2fc25c92c6e547029 WatchSource:0}: Error finding container 6651784c1f972e1750612313ea1ac3f179f2b73412516ec2fc25c92c6e547029: Status 404 returned error can't find the container with id 6651784c1f972e1750612313ea1ac3f179f2b73412516ec2fc25c92c6e547029 Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.408734 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc"] Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.988602 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" event={"ID":"a1cacabd-b318-4200-b357-181cb9bfdc78","Type":"ContainerStarted","Data":"a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149"} Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.989004 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" event={"ID":"a1cacabd-b318-4200-b357-181cb9bfdc78","Type":"ContainerStarted","Data":"6651784c1f972e1750612313ea1ac3f179f2b73412516ec2fc25c92c6e547029"} Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.989025 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.992868 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" event={"ID":"154fd6b3-3bd9-467d-9435-4417c448f831","Type":"ContainerStarted","Data":"fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6"} Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.992905 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" event={"ID":"154fd6b3-3bd9-467d-9435-4417c448f831","Type":"ContainerStarted","Data":"5d4e3b493c9d819b772d755baa983772890e045d816f644a6617b637f9b6f67c"} Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.993099 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:38 crc kubenswrapper[4838]: I1207 09:11:38.998127 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:39 crc kubenswrapper[4838]: I1207 09:11:39.001752 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:39 crc kubenswrapper[4838]: I1207 09:11:39.022391 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" podStartSLOduration=2.022375932 podStartE2EDuration="2.022375932s" podCreationTimestamp="2025-12-07 09:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:11:39.020638783 +0000 UTC m=+315.727957800" watchObservedRunningTime="2025-12-07 09:11:39.022375932 +0000 UTC m=+315.729694949" Dec 07 09:11:39 crc kubenswrapper[4838]: I1207 09:11:39.022850 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" podStartSLOduration=2.022844515 podStartE2EDuration="2.022844515s" podCreationTimestamp="2025-12-07 09:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:11:39.006073661 +0000 UTC m=+315.713392668" watchObservedRunningTime="2025-12-07 09:11:39.022844515 +0000 UTC m=+315.730163532" Dec 07 09:11:39 crc kubenswrapper[4838]: I1207 09:11:39.626328 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="950f31f5-e378-4a18-96f7-04e2091e2ca3" path="/var/lib/kubelet/pods/950f31f5-e378-4a18-96f7-04e2091e2ca3/volumes" Dec 07 09:11:39 crc kubenswrapper[4838]: I1207 09:11:39.628172 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b38416c-b5e5-4195-b950-6e449e035c8b" path="/var/lib/kubelet/pods/9b38416c-b5e5-4195-b950-6e449e035c8b/volumes" Dec 07 09:11:40 crc kubenswrapper[4838]: I1207 09:11:40.518358 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-lncht"] Dec 07 09:11:40 crc kubenswrapper[4838]: I1207 09:11:40.534436 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc"] Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.023242 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" podUID="a1cacabd-b318-4200-b357-181cb9bfdc78" containerName="controller-manager" containerID="cri-o://a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149" gracePeriod=30 Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.023361 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" podUID="154fd6b3-3bd9-467d-9435-4417c448f831" containerName="route-controller-manager" containerID="cri-o://fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6" gracePeriod=30 Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.475143 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.481258 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.510567 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt"] Dec 07 09:11:42 crc kubenswrapper[4838]: E1207 09:11:42.510832 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154fd6b3-3bd9-467d-9435-4417c448f831" containerName="route-controller-manager" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.510844 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="154fd6b3-3bd9-467d-9435-4417c448f831" containerName="route-controller-manager" Dec 07 09:11:42 crc kubenswrapper[4838]: E1207 09:11:42.510856 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1cacabd-b318-4200-b357-181cb9bfdc78" containerName="controller-manager" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.510864 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1cacabd-b318-4200-b357-181cb9bfdc78" containerName="controller-manager" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.510968 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1cacabd-b318-4200-b357-181cb9bfdc78" containerName="controller-manager" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.510983 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="154fd6b3-3bd9-467d-9435-4417c448f831" containerName="route-controller-manager" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.511377 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.519574 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt"] Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.600604 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-client-ca\") pod \"a1cacabd-b318-4200-b357-181cb9bfdc78\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.600951 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-config\") pod \"154fd6b3-3bd9-467d-9435-4417c448f831\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.600982 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-config\") pod \"a1cacabd-b318-4200-b357-181cb9bfdc78\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601002 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc5bl\" (UniqueName: \"kubernetes.io/projected/a1cacabd-b318-4200-b357-181cb9bfdc78-kube-api-access-qc5bl\") pod \"a1cacabd-b318-4200-b357-181cb9bfdc78\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601020 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmlcw\" (UniqueName: \"kubernetes.io/projected/154fd6b3-3bd9-467d-9435-4417c448f831-kube-api-access-jmlcw\") pod \"154fd6b3-3bd9-467d-9435-4417c448f831\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601041 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/154fd6b3-3bd9-467d-9435-4417c448f831-serving-cert\") pod \"154fd6b3-3bd9-467d-9435-4417c448f831\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601061 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-client-ca\") pod \"154fd6b3-3bd9-467d-9435-4417c448f831\" (UID: \"154fd6b3-3bd9-467d-9435-4417c448f831\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601110 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cacabd-b318-4200-b357-181cb9bfdc78-serving-cert\") pod \"a1cacabd-b318-4200-b357-181cb9bfdc78\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601136 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-proxy-ca-bundles\") pod \"a1cacabd-b318-4200-b357-181cb9bfdc78\" (UID: \"a1cacabd-b318-4200-b357-181cb9bfdc78\") " Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601258 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-config\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601333 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-client-ca\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601352 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7xm7\" (UniqueName: \"kubernetes.io/projected/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-kube-api-access-p7xm7\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601382 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-serving-cert\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601459 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-client-ca" (OuterVolumeSpecName: "client-ca") pod "a1cacabd-b318-4200-b357-181cb9bfdc78" (UID: "a1cacabd-b318-4200-b357-181cb9bfdc78"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.601881 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-config" (OuterVolumeSpecName: "config") pod "154fd6b3-3bd9-467d-9435-4417c448f831" (UID: "154fd6b3-3bd9-467d-9435-4417c448f831"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.602254 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a1cacabd-b318-4200-b357-181cb9bfdc78" (UID: "a1cacabd-b318-4200-b357-181cb9bfdc78"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.602612 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-client-ca" (OuterVolumeSpecName: "client-ca") pod "154fd6b3-3bd9-467d-9435-4417c448f831" (UID: "154fd6b3-3bd9-467d-9435-4417c448f831"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.602724 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-config" (OuterVolumeSpecName: "config") pod "a1cacabd-b318-4200-b357-181cb9bfdc78" (UID: "a1cacabd-b318-4200-b357-181cb9bfdc78"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.606327 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154fd6b3-3bd9-467d-9435-4417c448f831-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "154fd6b3-3bd9-467d-9435-4417c448f831" (UID: "154fd6b3-3bd9-467d-9435-4417c448f831"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.606393 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/154fd6b3-3bd9-467d-9435-4417c448f831-kube-api-access-jmlcw" (OuterVolumeSpecName: "kube-api-access-jmlcw") pod "154fd6b3-3bd9-467d-9435-4417c448f831" (UID: "154fd6b3-3bd9-467d-9435-4417c448f831"). InnerVolumeSpecName "kube-api-access-jmlcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.606601 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cacabd-b318-4200-b357-181cb9bfdc78-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a1cacabd-b318-4200-b357-181cb9bfdc78" (UID: "a1cacabd-b318-4200-b357-181cb9bfdc78"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.607098 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1cacabd-b318-4200-b357-181cb9bfdc78-kube-api-access-qc5bl" (OuterVolumeSpecName: "kube-api-access-qc5bl") pod "a1cacabd-b318-4200-b357-181cb9bfdc78" (UID: "a1cacabd-b318-4200-b357-181cb9bfdc78"). InnerVolumeSpecName "kube-api-access-qc5bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.702804 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-client-ca\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.702899 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7xm7\" (UniqueName: \"kubernetes.io/projected/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-kube-api-access-p7xm7\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.702945 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-serving-cert\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703006 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-config\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703074 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703090 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cacabd-b318-4200-b357-181cb9bfdc78-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703102 4838 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703116 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703128 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/154fd6b3-3bd9-467d-9435-4417c448f831-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703139 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cacabd-b318-4200-b357-181cb9bfdc78-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703151 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc5bl\" (UniqueName: \"kubernetes.io/projected/a1cacabd-b318-4200-b357-181cb9bfdc78-kube-api-access-qc5bl\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703166 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmlcw\" (UniqueName: \"kubernetes.io/projected/154fd6b3-3bd9-467d-9435-4417c448f831-kube-api-access-jmlcw\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.703178 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/154fd6b3-3bd9-467d-9435-4417c448f831-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.704939 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-config\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.705888 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-client-ca\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.709293 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-serving-cert\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.721253 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7xm7\" (UniqueName: \"kubernetes.io/projected/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-kube-api-access-p7xm7\") pod \"route-controller-manager-5fdfb79994-tbsjt\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:42 crc kubenswrapper[4838]: I1207 09:11:42.829572 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.040194 4838 generic.go:334] "Generic (PLEG): container finished" podID="154fd6b3-3bd9-467d-9435-4417c448f831" containerID="fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6" exitCode=0 Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.040242 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.040256 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" event={"ID":"154fd6b3-3bd9-467d-9435-4417c448f831","Type":"ContainerDied","Data":"fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6"} Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.040658 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc" event={"ID":"154fd6b3-3bd9-467d-9435-4417c448f831","Type":"ContainerDied","Data":"5d4e3b493c9d819b772d755baa983772890e045d816f644a6617b637f9b6f67c"} Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.040677 4838 scope.go:117] "RemoveContainer" containerID="fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.044449 4838 generic.go:334] "Generic (PLEG): container finished" podID="a1cacabd-b318-4200-b357-181cb9bfdc78" containerID="a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149" exitCode=0 Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.044491 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" event={"ID":"a1cacabd-b318-4200-b357-181cb9bfdc78","Type":"ContainerDied","Data":"a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149"} Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.044511 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" event={"ID":"a1cacabd-b318-4200-b357-181cb9bfdc78","Type":"ContainerDied","Data":"6651784c1f972e1750612313ea1ac3f179f2b73412516ec2fc25c92c6e547029"} Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.044535 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df5f84b58-lncht" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.066392 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt"] Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.075281 4838 scope.go:117] "RemoveContainer" containerID="fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.076184 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc"] Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.080435 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-wp6qc"] Dec 07 09:11:43 crc kubenswrapper[4838]: E1207 09:11:43.083400 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6\": container with ID starting with fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6 not found: ID does not exist" containerID="fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.083477 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6"} err="failed to get container status \"fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6\": rpc error: code = NotFound desc = could not find container \"fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6\": container with ID starting with fdd4130c958c6b78e7486b6b5afb1192c91652ed0bf0ec158f654ed0d27cf1e6 not found: ID does not exist" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.083514 4838 scope.go:117] "RemoveContainer" containerID="a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.090841 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-lncht"] Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.099558 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-lncht"] Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.115041 4838 scope.go:117] "RemoveContainer" containerID="a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149" Dec 07 09:11:43 crc kubenswrapper[4838]: E1207 09:11:43.115912 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149\": container with ID starting with a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149 not found: ID does not exist" containerID="a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.116218 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149"} err="failed to get container status \"a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149\": rpc error: code = NotFound desc = could not find container \"a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149\": container with ID starting with a957e7b7c72d9d788e026d196841d65ade13a844b5dfc8aaa902f8c519364149 not found: ID does not exist" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.620783 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="154fd6b3-3bd9-467d-9435-4417c448f831" path="/var/lib/kubelet/pods/154fd6b3-3bd9-467d-9435-4417c448f831/volumes" Dec 07 09:11:43 crc kubenswrapper[4838]: I1207 09:11:43.621594 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1cacabd-b318-4200-b357-181cb9bfdc78" path="/var/lib/kubelet/pods/a1cacabd-b318-4200-b357-181cb9bfdc78/volumes" Dec 07 09:11:44 crc kubenswrapper[4838]: I1207 09:11:44.052882 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" event={"ID":"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8","Type":"ContainerStarted","Data":"895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd"} Dec 07 09:11:44 crc kubenswrapper[4838]: I1207 09:11:44.052918 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" event={"ID":"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8","Type":"ContainerStarted","Data":"24a8c07cfea0e77599ff7a2c133d3f2db3be8253dcf3e544237c621f0c4e8928"} Dec 07 09:11:44 crc kubenswrapper[4838]: I1207 09:11:44.053099 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:44 crc kubenswrapper[4838]: I1207 09:11:44.057865 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:11:44 crc kubenswrapper[4838]: I1207 09:11:44.084588 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" podStartSLOduration=4.084571196 podStartE2EDuration="4.084571196s" podCreationTimestamp="2025-12-07 09:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:11:44.083481715 +0000 UTC m=+320.790800732" watchObservedRunningTime="2025-12-07 09:11:44.084571196 +0000 UTC m=+320.791890223" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.060057 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-kmpbw"] Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.061719 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.065389 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.065521 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.066308 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.066385 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.066978 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.067923 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.080150 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.089071 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-kmpbw"] Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.137344 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-client-ca\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.137397 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49e5774f-270f-4d72-9b50-8e6789371636-serving-cert\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.137448 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfdsv\" (UniqueName: \"kubernetes.io/projected/49e5774f-270f-4d72-9b50-8e6789371636-kube-api-access-qfdsv\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.137556 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-config\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.137630 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.238947 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-client-ca\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.238991 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49e5774f-270f-4d72-9b50-8e6789371636-serving-cert\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.239028 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfdsv\" (UniqueName: \"kubernetes.io/projected/49e5774f-270f-4d72-9b50-8e6789371636-kube-api-access-qfdsv\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.239048 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-config\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.239069 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.239973 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-client-ca\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.240238 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.240532 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-config\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.255021 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49e5774f-270f-4d72-9b50-8e6789371636-serving-cert\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.263757 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfdsv\" (UniqueName: \"kubernetes.io/projected/49e5774f-270f-4d72-9b50-8e6789371636-kube-api-access-qfdsv\") pod \"controller-manager-589f9f9656-kmpbw\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.398364 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:45 crc kubenswrapper[4838]: I1207 09:11:45.683586 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-kmpbw"] Dec 07 09:11:46 crc kubenswrapper[4838]: I1207 09:11:46.066678 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" event={"ID":"49e5774f-270f-4d72-9b50-8e6789371636","Type":"ContainerStarted","Data":"062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6"} Dec 07 09:11:46 crc kubenswrapper[4838]: I1207 09:11:46.067000 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" event={"ID":"49e5774f-270f-4d72-9b50-8e6789371636","Type":"ContainerStarted","Data":"d1fbc5c384fd4fe4ed568f7daf05896cc2ca79f10c6e27bd7af663103c23f449"} Dec 07 09:11:46 crc kubenswrapper[4838]: I1207 09:11:46.068573 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:46 crc kubenswrapper[4838]: I1207 09:11:46.077544 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:11:46 crc kubenswrapper[4838]: I1207 09:11:46.086603 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" podStartSLOduration=6.08658813 podStartE2EDuration="6.08658813s" podCreationTimestamp="2025-12-07 09:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:11:46.085963572 +0000 UTC m=+322.793282599" watchObservedRunningTime="2025-12-07 09:11:46.08658813 +0000 UTC m=+322.793907147" Dec 07 09:11:54 crc kubenswrapper[4838]: I1207 09:11:54.493455 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:11:54 crc kubenswrapper[4838]: I1207 09:11:54.494146 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:12:16 crc kubenswrapper[4838]: I1207 09:12:16.703137 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt"] Dec 07 09:12:16 crc kubenswrapper[4838]: I1207 09:12:16.703910 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" podUID="d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" containerName="route-controller-manager" containerID="cri-o://895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd" gracePeriod=30 Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.165255 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.256919 4838 generic.go:334] "Generic (PLEG): container finished" podID="d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" containerID="895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd" exitCode=0 Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.256976 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" event={"ID":"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8","Type":"ContainerDied","Data":"895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd"} Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.257006 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" event={"ID":"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8","Type":"ContainerDied","Data":"24a8c07cfea0e77599ff7a2c133d3f2db3be8253dcf3e544237c621f0c4e8928"} Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.257026 4838 scope.go:117] "RemoveContainer" containerID="895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.257067 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.272955 4838 scope.go:117] "RemoveContainer" containerID="895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd" Dec 07 09:12:17 crc kubenswrapper[4838]: E1207 09:12:17.273863 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd\": container with ID starting with 895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd not found: ID does not exist" containerID="895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.274105 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd"} err="failed to get container status \"895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd\": rpc error: code = NotFound desc = could not find container \"895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd\": container with ID starting with 895e8ae7569d59f27373d49be57bbeaecbbb757c4469ba2f0b7b96c11848bafd not found: ID does not exist" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.286176 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-config\") pod \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.286278 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7xm7\" (UniqueName: \"kubernetes.io/projected/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-kube-api-access-p7xm7\") pod \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.286333 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-serving-cert\") pod \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.286453 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-client-ca\") pod \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\" (UID: \"d63fce7e-34bb-49a7-9ed4-81d0c60f79a8\") " Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.287187 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-config" (OuterVolumeSpecName: "config") pod "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" (UID: "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.287415 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-client-ca" (OuterVolumeSpecName: "client-ca") pod "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" (UID: "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.292066 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" (UID: "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.292726 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-kube-api-access-p7xm7" (OuterVolumeSpecName: "kube-api-access-p7xm7") pod "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" (UID: "d63fce7e-34bb-49a7-9ed4-81d0c60f79a8"). InnerVolumeSpecName "kube-api-access-p7xm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.388158 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.388221 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.388233 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7xm7\" (UniqueName: \"kubernetes.io/projected/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-kube-api-access-p7xm7\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.388243 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.580667 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt"] Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.585730 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-tbsjt"] Dec 07 09:12:17 crc kubenswrapper[4838]: I1207 09:12:17.621225 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" path="/var/lib/kubelet/pods/d63fce7e-34bb-49a7-9ed4-81d0c60f79a8/volumes" Dec 07 09:12:17 crc kubenswrapper[4838]: E1207 09:12:17.638015 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd63fce7e_34bb_49a7_9ed4_81d0c60f79a8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd63fce7e_34bb_49a7_9ed4_81d0c60f79a8.slice/crio-24a8c07cfea0e77599ff7a2c133d3f2db3be8253dcf3e544237c621f0c4e8928\": RecentStats: unable to find data in memory cache]" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.081949 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6"] Dec 07 09:12:18 crc kubenswrapper[4838]: E1207 09:12:18.082716 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" containerName="route-controller-manager" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.082732 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" containerName="route-controller-manager" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.082869 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63fce7e-34bb-49a7-9ed4-81d0c60f79a8" containerName="route-controller-manager" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.083365 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.086872 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.087527 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.090390 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.090417 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.090486 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.091371 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.094136 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6"] Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.197243 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-config\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.197290 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-serving-cert\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.197319 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-client-ca\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.197347 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czrng\" (UniqueName: \"kubernetes.io/projected/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-kube-api-access-czrng\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.298666 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-config\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.299018 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-serving-cert\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.300003 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-client-ca\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.300240 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czrng\" (UniqueName: \"kubernetes.io/projected/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-kube-api-access-czrng\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.300489 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-config\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.302483 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-client-ca\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.303371 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-serving-cert\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.323728 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czrng\" (UniqueName: \"kubernetes.io/projected/c870d4e5-aeeb-49b8-a238-7dc6c06f76de-kube-api-access-czrng\") pod \"route-controller-manager-cf78f955c-mzjz6\" (UID: \"c870d4e5-aeeb-49b8-a238-7dc6c06f76de\") " pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.452689 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:18 crc kubenswrapper[4838]: I1207 09:12:18.862720 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6"] Dec 07 09:12:19 crc kubenswrapper[4838]: I1207 09:12:19.266622 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" event={"ID":"c870d4e5-aeeb-49b8-a238-7dc6c06f76de","Type":"ContainerStarted","Data":"ebafb2e922d098cbe86d4cbcdcee95d572711d9a71296da18b7179cc7b061316"} Dec 07 09:12:19 crc kubenswrapper[4838]: I1207 09:12:19.266674 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" event={"ID":"c870d4e5-aeeb-49b8-a238-7dc6c06f76de","Type":"ContainerStarted","Data":"299747a5619aa51d91b4f6bbe3f275a83fb98c08baf7ae7b2d8bd36e0419847c"} Dec 07 09:12:19 crc kubenswrapper[4838]: I1207 09:12:19.267861 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:19 crc kubenswrapper[4838]: I1207 09:12:19.289145 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" podStartSLOduration=3.289125931 podStartE2EDuration="3.289125931s" podCreationTimestamp="2025-12-07 09:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:12:19.283199558 +0000 UTC m=+355.990518585" watchObservedRunningTime="2025-12-07 09:12:19.289125931 +0000 UTC m=+355.996444948" Dec 07 09:12:19 crc kubenswrapper[4838]: I1207 09:12:19.373747 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cf78f955c-mzjz6" Dec 07 09:12:24 crc kubenswrapper[4838]: I1207 09:12:24.492864 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:12:24 crc kubenswrapper[4838]: I1207 09:12:24.493215 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:12:33 crc kubenswrapper[4838]: I1207 09:12:33.808115 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hk46c"] Dec 07 09:12:33 crc kubenswrapper[4838]: I1207 09:12:33.809465 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:33 crc kubenswrapper[4838]: I1207 09:12:33.837711 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hk46c"] Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.003522 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97de1ddf-bbfd-482c-8197-143df1c39fe4-trusted-ca\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.003803 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97de1ddf-bbfd-482c-8197-143df1c39fe4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.003939 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-bound-sa-token\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.004022 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.004111 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97de1ddf-bbfd-482c-8197-143df1c39fe4-registry-certificates\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.004218 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dp4c\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-kube-api-access-5dp4c\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.004342 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97de1ddf-bbfd-482c-8197-143df1c39fe4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.004463 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-registry-tls\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.027767 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.105222 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-registry-tls\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.105483 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97de1ddf-bbfd-482c-8197-143df1c39fe4-trusted-ca\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.105565 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97de1ddf-bbfd-482c-8197-143df1c39fe4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.105656 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-bound-sa-token\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.105809 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97de1ddf-bbfd-482c-8197-143df1c39fe4-registry-certificates\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.105927 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dp4c\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-kube-api-access-5dp4c\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.106009 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97de1ddf-bbfd-482c-8197-143df1c39fe4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.106462 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97de1ddf-bbfd-482c-8197-143df1c39fe4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.108432 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97de1ddf-bbfd-482c-8197-143df1c39fe4-trusted-ca\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.108554 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97de1ddf-bbfd-482c-8197-143df1c39fe4-registry-certificates\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.117088 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97de1ddf-bbfd-482c-8197-143df1c39fe4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.120275 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-registry-tls\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.122227 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dp4c\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-kube-api-access-5dp4c\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.122243 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97de1ddf-bbfd-482c-8197-143df1c39fe4-bound-sa-token\") pod \"image-registry-66df7c8f76-hk46c\" (UID: \"97de1ddf-bbfd-482c-8197-143df1c39fe4\") " pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.125729 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:34 crc kubenswrapper[4838]: I1207 09:12:34.543729 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hk46c"] Dec 07 09:12:35 crc kubenswrapper[4838]: I1207 09:12:35.361974 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" event={"ID":"97de1ddf-bbfd-482c-8197-143df1c39fe4","Type":"ContainerStarted","Data":"b8b0c96af36a420dfd18b01cc0ef20c3287434e6e7c4c9821518c038f1521fd3"} Dec 07 09:12:35 crc kubenswrapper[4838]: I1207 09:12:35.362013 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" event={"ID":"97de1ddf-bbfd-482c-8197-143df1c39fe4","Type":"ContainerStarted","Data":"70102a4ed849c81fcd9852e62c7bf83771ba425e1721381f4df746a17cb5e8bb"} Dec 07 09:12:35 crc kubenswrapper[4838]: I1207 09:12:35.363053 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:35 crc kubenswrapper[4838]: I1207 09:12:35.398693 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" podStartSLOduration=2.398663419 podStartE2EDuration="2.398663419s" podCreationTimestamp="2025-12-07 09:12:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:12:35.39149466 +0000 UTC m=+372.098813677" watchObservedRunningTime="2025-12-07 09:12:35.398663419 +0000 UTC m=+372.105982476" Dec 07 09:12:36 crc kubenswrapper[4838]: I1207 09:12:36.724039 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-kmpbw"] Dec 07 09:12:36 crc kubenswrapper[4838]: I1207 09:12:36.724460 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" podUID="49e5774f-270f-4d72-9b50-8e6789371636" containerName="controller-manager" containerID="cri-o://062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6" gracePeriod=30 Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.096675 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.265201 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-config\") pod \"49e5774f-270f-4d72-9b50-8e6789371636\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.265269 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfdsv\" (UniqueName: \"kubernetes.io/projected/49e5774f-270f-4d72-9b50-8e6789371636-kube-api-access-qfdsv\") pod \"49e5774f-270f-4d72-9b50-8e6789371636\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.265310 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49e5774f-270f-4d72-9b50-8e6789371636-serving-cert\") pod \"49e5774f-270f-4d72-9b50-8e6789371636\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.265328 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-proxy-ca-bundles\") pod \"49e5774f-270f-4d72-9b50-8e6789371636\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.265342 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-client-ca\") pod \"49e5774f-270f-4d72-9b50-8e6789371636\" (UID: \"49e5774f-270f-4d72-9b50-8e6789371636\") " Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.266172 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "49e5774f-270f-4d72-9b50-8e6789371636" (UID: "49e5774f-270f-4d72-9b50-8e6789371636"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.266298 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-config" (OuterVolumeSpecName: "config") pod "49e5774f-270f-4d72-9b50-8e6789371636" (UID: "49e5774f-270f-4d72-9b50-8e6789371636"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.266337 4838 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.266664 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-client-ca" (OuterVolumeSpecName: "client-ca") pod "49e5774f-270f-4d72-9b50-8e6789371636" (UID: "49e5774f-270f-4d72-9b50-8e6789371636"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.276013 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e5774f-270f-4d72-9b50-8e6789371636-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "49e5774f-270f-4d72-9b50-8e6789371636" (UID: "49e5774f-270f-4d72-9b50-8e6789371636"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.283029 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e5774f-270f-4d72-9b50-8e6789371636-kube-api-access-qfdsv" (OuterVolumeSpecName: "kube-api-access-qfdsv") pod "49e5774f-270f-4d72-9b50-8e6789371636" (UID: "49e5774f-270f-4d72-9b50-8e6789371636"). InnerVolumeSpecName "kube-api-access-qfdsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.367763 4838 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49e5774f-270f-4d72-9b50-8e6789371636-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.367796 4838 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.367806 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5774f-270f-4d72-9b50-8e6789371636-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.367846 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfdsv\" (UniqueName: \"kubernetes.io/projected/49e5774f-270f-4d72-9b50-8e6789371636-kube-api-access-qfdsv\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.375664 4838 generic.go:334] "Generic (PLEG): container finished" podID="49e5774f-270f-4d72-9b50-8e6789371636" containerID="062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6" exitCode=0 Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.375703 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" event={"ID":"49e5774f-270f-4d72-9b50-8e6789371636","Type":"ContainerDied","Data":"062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6"} Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.375728 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" event={"ID":"49e5774f-270f-4d72-9b50-8e6789371636","Type":"ContainerDied","Data":"d1fbc5c384fd4fe4ed568f7daf05896cc2ca79f10c6e27bd7af663103c23f449"} Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.375744 4838 scope.go:117] "RemoveContainer" containerID="062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.375854 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-kmpbw" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.398460 4838 scope.go:117] "RemoveContainer" containerID="062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6" Dec 07 09:12:37 crc kubenswrapper[4838]: E1207 09:12:37.399230 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6\": container with ID starting with 062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6 not found: ID does not exist" containerID="062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.399267 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6"} err="failed to get container status \"062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6\": rpc error: code = NotFound desc = could not find container \"062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6\": container with ID starting with 062dfad5c9d86458379c958a9afcfd1b1ff914fadb0f6c45cc399372bee76ec6 not found: ID does not exist" Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.406500 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-kmpbw"] Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.414354 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-kmpbw"] Dec 07 09:12:37 crc kubenswrapper[4838]: I1207 09:12:37.626436 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e5774f-270f-4d72-9b50-8e6789371636" path="/var/lib/kubelet/pods/49e5774f-270f-4d72-9b50-8e6789371636/volumes" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.100152 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-2lfqj"] Dec 07 09:12:38 crc kubenswrapper[4838]: E1207 09:12:38.100806 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e5774f-270f-4d72-9b50-8e6789371636" containerName="controller-manager" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.100848 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e5774f-270f-4d72-9b50-8e6789371636" containerName="controller-manager" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.100995 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e5774f-270f-4d72-9b50-8e6789371636" containerName="controller-manager" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.101558 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.107464 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.112863 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.113129 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.113239 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.113470 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-2lfqj"] Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.113496 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.114468 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.121911 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.279733 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f6r2\" (UniqueName: \"kubernetes.io/projected/71ea38f7-dbcd-4444-bdd9-ab896b20e360-kube-api-access-7f6r2\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.279806 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-proxy-ca-bundles\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.280037 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-config\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.280075 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-client-ca\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.280226 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71ea38f7-dbcd-4444-bdd9-ab896b20e360-serving-cert\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.381185 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71ea38f7-dbcd-4444-bdd9-ab896b20e360-serving-cert\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.381245 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f6r2\" (UniqueName: \"kubernetes.io/projected/71ea38f7-dbcd-4444-bdd9-ab896b20e360-kube-api-access-7f6r2\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.381270 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-proxy-ca-bundles\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.381330 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-config\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.381351 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-client-ca\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.382396 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-client-ca\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.387585 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-proxy-ca-bundles\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.387783 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71ea38f7-dbcd-4444-bdd9-ab896b20e360-config\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.393564 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71ea38f7-dbcd-4444-bdd9-ab896b20e360-serving-cert\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.412148 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f6r2\" (UniqueName: \"kubernetes.io/projected/71ea38f7-dbcd-4444-bdd9-ab896b20e360-kube-api-access-7f6r2\") pod \"controller-manager-6df5f84b58-2lfqj\" (UID: \"71ea38f7-dbcd-4444-bdd9-ab896b20e360\") " pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.432086 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:38 crc kubenswrapper[4838]: I1207 09:12:38.656709 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6df5f84b58-2lfqj"] Dec 07 09:12:38 crc kubenswrapper[4838]: W1207 09:12:38.668481 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71ea38f7_dbcd_4444_bdd9_ab896b20e360.slice/crio-4678c5b4b09b19bdc44c81858e7bd97734cfb9c15b7545e6437a44d4e856553c WatchSource:0}: Error finding container 4678c5b4b09b19bdc44c81858e7bd97734cfb9c15b7545e6437a44d4e856553c: Status 404 returned error can't find the container with id 4678c5b4b09b19bdc44c81858e7bd97734cfb9c15b7545e6437a44d4e856553c Dec 07 09:12:39 crc kubenswrapper[4838]: I1207 09:12:39.396997 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" event={"ID":"71ea38f7-dbcd-4444-bdd9-ab896b20e360","Type":"ContainerStarted","Data":"efbee024f3c5481337e461663237108655737a55f59ffe155f3e976715f336a3"} Dec 07 09:12:39 crc kubenswrapper[4838]: I1207 09:12:39.397044 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" event={"ID":"71ea38f7-dbcd-4444-bdd9-ab896b20e360","Type":"ContainerStarted","Data":"4678c5b4b09b19bdc44c81858e7bd97734cfb9c15b7545e6437a44d4e856553c"} Dec 07 09:12:39 crc kubenswrapper[4838]: I1207 09:12:39.397339 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:39 crc kubenswrapper[4838]: I1207 09:12:39.405831 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" Dec 07 09:12:39 crc kubenswrapper[4838]: I1207 09:12:39.439501 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6df5f84b58-2lfqj" podStartSLOduration=3.439486573 podStartE2EDuration="3.439486573s" podCreationTimestamp="2025-12-07 09:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:12:39.420371446 +0000 UTC m=+376.127690473" watchObservedRunningTime="2025-12-07 09:12:39.439486573 +0000 UTC m=+376.146805590" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.695754 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fh5rj"] Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.696567 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fh5rj" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="registry-server" containerID="cri-o://4f6b7bb02c589b7341245c2556d6dd34da38cb6169f3849ede8b4f89f0aff990" gracePeriod=30 Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.713621 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4z85d"] Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.713914 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4z85d" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="registry-server" containerID="cri-o://5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa" gracePeriod=30 Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.723984 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll54w"] Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.724234 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" containerID="cri-o://3ec4f3c88edacf90710520d8ed393bdf2fcaa4a691fc0048e77f68fb9a2a37d3" gracePeriod=30 Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.728395 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kn79"] Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.728650 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7kn79" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="registry-server" containerID="cri-o://9f30edaf5cf0fc6cc6b7c4e4bf076e0e2f44b2d156f4fcb792db41d85e062071" gracePeriod=30 Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.739253 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d58v2"] Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.739515 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d58v2" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="registry-server" containerID="cri-o://cfe1382645341fe7929cd53ccc106774d41ff28adb85f0b5541b9916a9900dc1" gracePeriod=30 Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.752519 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-69xn8"] Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.753398 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.769451 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-69xn8"] Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.886497 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsnxj\" (UniqueName: \"kubernetes.io/projected/c21136ea-de51-48f3-b79f-493f2d88ece5-kube-api-access-xsnxj\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.886752 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c21136ea-de51-48f3-b79f-493f2d88ece5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.886839 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c21136ea-de51-48f3-b79f-493f2d88ece5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.987655 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c21136ea-de51-48f3-b79f-493f2d88ece5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.987772 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsnxj\" (UniqueName: \"kubernetes.io/projected/c21136ea-de51-48f3-b79f-493f2d88ece5-kube-api-access-xsnxj\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.987802 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c21136ea-de51-48f3-b79f-493f2d88ece5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.988899 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c21136ea-de51-48f3-b79f-493f2d88ece5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:44 crc kubenswrapper[4838]: I1207 09:12:44.994581 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c21136ea-de51-48f3-b79f-493f2d88ece5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.034608 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsnxj\" (UniqueName: \"kubernetes.io/projected/c21136ea-de51-48f3-b79f-493f2d88ece5-kube-api-access-xsnxj\") pod \"marketplace-operator-79b997595-69xn8\" (UID: \"c21136ea-de51-48f3-b79f-493f2d88ece5\") " pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.074664 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.339068 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.430171 4838 generic.go:334] "Generic (PLEG): container finished" podID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerID="5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa" exitCode=0 Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.430226 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4z85d" event={"ID":"e3208dc2-6618-440a-816c-cab7c4fc394c","Type":"ContainerDied","Data":"5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa"} Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.430254 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4z85d" event={"ID":"e3208dc2-6618-440a-816c-cab7c4fc394c","Type":"ContainerDied","Data":"faceeafa7d1285f7176c1343ca208a8a3b3c46f94e1f54caec1f28dad4e32ef7"} Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.430269 4838 scope.go:117] "RemoveContainer" containerID="5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.430373 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4z85d" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.461209 4838 generic.go:334] "Generic (PLEG): container finished" podID="be195d3f-68b9-47a0-b726-6e7736450daa" containerID="9f30edaf5cf0fc6cc6b7c4e4bf076e0e2f44b2d156f4fcb792db41d85e062071" exitCode=0 Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.461283 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kn79" event={"ID":"be195d3f-68b9-47a0-b726-6e7736450daa","Type":"ContainerDied","Data":"9f30edaf5cf0fc6cc6b7c4e4bf076e0e2f44b2d156f4fcb792db41d85e062071"} Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.478246 4838 generic.go:334] "Generic (PLEG): container finished" podID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerID="cfe1382645341fe7929cd53ccc106774d41ff28adb85f0b5541b9916a9900dc1" exitCode=0 Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.478307 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d58v2" event={"ID":"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3","Type":"ContainerDied","Data":"cfe1382645341fe7929cd53ccc106774d41ff28adb85f0b5541b9916a9900dc1"} Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.479743 4838 scope.go:117] "RemoveContainer" containerID="47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.481539 4838 generic.go:334] "Generic (PLEG): container finished" podID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerID="4f6b7bb02c589b7341245c2556d6dd34da38cb6169f3849ede8b4f89f0aff990" exitCode=0 Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.481580 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fh5rj" event={"ID":"26766408-4baa-43e9-9a1d-9b0af9fffe61","Type":"ContainerDied","Data":"4f6b7bb02c589b7341245c2556d6dd34da38cb6169f3849ede8b4f89f0aff990"} Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.490413 4838 generic.go:334] "Generic (PLEG): container finished" podID="3854bccd-8788-4cce-8a17-8d06939f172b" containerID="3ec4f3c88edacf90710520d8ed393bdf2fcaa4a691fc0048e77f68fb9a2a37d3" exitCode=0 Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.490447 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" event={"ID":"3854bccd-8788-4cce-8a17-8d06939f172b","Type":"ContainerDied","Data":"3ec4f3c88edacf90710520d8ed393bdf2fcaa4a691fc0048e77f68fb9a2a37d3"} Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.500045 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-catalog-content\") pod \"e3208dc2-6618-440a-816c-cab7c4fc394c\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.500618 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-utilities\") pod \"e3208dc2-6618-440a-816c-cab7c4fc394c\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.500742 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdhsd\" (UniqueName: \"kubernetes.io/projected/e3208dc2-6618-440a-816c-cab7c4fc394c-kube-api-access-cdhsd\") pod \"e3208dc2-6618-440a-816c-cab7c4fc394c\" (UID: \"e3208dc2-6618-440a-816c-cab7c4fc394c\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.502225 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-utilities" (OuterVolumeSpecName: "utilities") pod "e3208dc2-6618-440a-816c-cab7c4fc394c" (UID: "e3208dc2-6618-440a-816c-cab7c4fc394c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.505898 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3208dc2-6618-440a-816c-cab7c4fc394c-kube-api-access-cdhsd" (OuterVolumeSpecName: "kube-api-access-cdhsd") pod "e3208dc2-6618-440a-816c-cab7c4fc394c" (UID: "e3208dc2-6618-440a-816c-cab7c4fc394c"). InnerVolumeSpecName "kube-api-access-cdhsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.519699 4838 scope.go:117] "RemoveContainer" containerID="46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.546347 4838 scope.go:117] "RemoveContainer" containerID="5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa" Dec 07 09:12:45 crc kubenswrapper[4838]: E1207 09:12:45.546884 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa\": container with ID starting with 5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa not found: ID does not exist" containerID="5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.546956 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa"} err="failed to get container status \"5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa\": rpc error: code = NotFound desc = could not find container \"5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa\": container with ID starting with 5b7f025f5e516cafa71ca50e5187e2623839a97862e839963c4a130713ab31fa not found: ID does not exist" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.546990 4838 scope.go:117] "RemoveContainer" containerID="47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e" Dec 07 09:12:45 crc kubenswrapper[4838]: E1207 09:12:45.547471 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e\": container with ID starting with 47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e not found: ID does not exist" containerID="47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.547537 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e"} err="failed to get container status \"47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e\": rpc error: code = NotFound desc = could not find container \"47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e\": container with ID starting with 47c1c8945764319ac0ed0b1ff122e5093c552b0a01e179ab8059bf3714ef664e not found: ID does not exist" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.547570 4838 scope.go:117] "RemoveContainer" containerID="46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945" Dec 07 09:12:45 crc kubenswrapper[4838]: E1207 09:12:45.548096 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945\": container with ID starting with 46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945 not found: ID does not exist" containerID="46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.548130 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945"} err="failed to get container status \"46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945\": rpc error: code = NotFound desc = could not find container \"46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945\": container with ID starting with 46b8a5ce93a09083066b214ae4c1f1fe74d695cbaebabca982f2a950004a9945 not found: ID does not exist" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.548157 4838 scope.go:117] "RemoveContainer" containerID="948a67c69d5c4326b50d2b256fc103967ff572d7d019d4a2bfa3cdafb23f1a38" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.559394 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.566316 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.588227 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.596695 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3208dc2-6618-440a-816c-cab7c4fc394c" (UID: "e3208dc2-6618-440a-816c-cab7c4fc394c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.605758 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca\") pod \"3854bccd-8788-4cce-8a17-8d06939f172b\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.606055 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ksbk\" (UniqueName: \"kubernetes.io/projected/be195d3f-68b9-47a0-b726-6e7736450daa-kube-api-access-6ksbk\") pod \"be195d3f-68b9-47a0-b726-6e7736450daa\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.606097 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp9c2\" (UniqueName: \"kubernetes.io/projected/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-kube-api-access-fp9c2\") pod \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.606125 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics\") pod \"3854bccd-8788-4cce-8a17-8d06939f172b\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.606358 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdhsd\" (UniqueName: \"kubernetes.io/projected/e3208dc2-6618-440a-816c-cab7c4fc394c-kube-api-access-cdhsd\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.606370 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.606379 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3208dc2-6618-440a-816c-cab7c4fc394c-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.606625 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "3854bccd-8788-4cce-8a17-8d06939f172b" (UID: "3854bccd-8788-4cce-8a17-8d06939f172b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.609966 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be195d3f-68b9-47a0-b726-6e7736450daa-kube-api-access-6ksbk" (OuterVolumeSpecName: "kube-api-access-6ksbk") pod "be195d3f-68b9-47a0-b726-6e7736450daa" (UID: "be195d3f-68b9-47a0-b726-6e7736450daa"). InnerVolumeSpecName "kube-api-access-6ksbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.611662 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "3854bccd-8788-4cce-8a17-8d06939f172b" (UID: "3854bccd-8788-4cce-8a17-8d06939f172b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.622665 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-kube-api-access-fp9c2" (OuterVolumeSpecName: "kube-api-access-fp9c2") pod "b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" (UID: "b3ab327e-67a8-4c75-809c-9ecb0d4a9be3"). InnerVolumeSpecName "kube-api-access-fp9c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.706707 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-utilities\") pod \"be195d3f-68b9-47a0-b726-6e7736450daa\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.706884 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv6kn\" (UniqueName: \"kubernetes.io/projected/3854bccd-8788-4cce-8a17-8d06939f172b-kube-api-access-lv6kn\") pod \"3854bccd-8788-4cce-8a17-8d06939f172b\" (UID: \"3854bccd-8788-4cce-8a17-8d06939f172b\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707336 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-utilities\") pod \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707398 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-catalog-content\") pod \"be195d3f-68b9-47a0-b726-6e7736450daa\" (UID: \"be195d3f-68b9-47a0-b726-6e7736450daa\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707415 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-catalog-content\") pod \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\" (UID: \"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707406 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-utilities" (OuterVolumeSpecName: "utilities") pod "be195d3f-68b9-47a0-b726-6e7736450daa" (UID: "be195d3f-68b9-47a0-b726-6e7736450daa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707676 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp9c2\" (UniqueName: \"kubernetes.io/projected/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-kube-api-access-fp9c2\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707692 4838 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707702 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707711 4838 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3854bccd-8788-4cce-8a17-8d06939f172b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.707719 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ksbk\" (UniqueName: \"kubernetes.io/projected/be195d3f-68b9-47a0-b726-6e7736450daa-kube-api-access-6ksbk\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.708228 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-utilities" (OuterVolumeSpecName: "utilities") pod "b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" (UID: "b3ab327e-67a8-4c75-809c-9ecb0d4a9be3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.709727 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3854bccd-8788-4cce-8a17-8d06939f172b-kube-api-access-lv6kn" (OuterVolumeSpecName: "kube-api-access-lv6kn") pod "3854bccd-8788-4cce-8a17-8d06939f172b" (UID: "3854bccd-8788-4cce-8a17-8d06939f172b"). InnerVolumeSpecName "kube-api-access-lv6kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.715898 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-69xn8"] Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.732141 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be195d3f-68b9-47a0-b726-6e7736450daa" (UID: "be195d3f-68b9-47a0-b726-6e7736450daa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.759750 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4z85d"] Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.760768 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.778643 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4z85d"] Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.814177 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be195d3f-68b9-47a0-b726-6e7736450daa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.815330 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv6kn\" (UniqueName: \"kubernetes.io/projected/3854bccd-8788-4cce-8a17-8d06939f172b-kube-api-access-lv6kn\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.815349 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.888014 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" (UID: "b3ab327e-67a8-4c75-809c-9ecb0d4a9be3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.915868 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxbj6\" (UniqueName: \"kubernetes.io/projected/26766408-4baa-43e9-9a1d-9b0af9fffe61-kube-api-access-sxbj6\") pod \"26766408-4baa-43e9-9a1d-9b0af9fffe61\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.915998 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-utilities\") pod \"26766408-4baa-43e9-9a1d-9b0af9fffe61\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.916930 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-utilities" (OuterVolumeSpecName: "utilities") pod "26766408-4baa-43e9-9a1d-9b0af9fffe61" (UID: "26766408-4baa-43e9-9a1d-9b0af9fffe61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.916994 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-catalog-content\") pod \"26766408-4baa-43e9-9a1d-9b0af9fffe61\" (UID: \"26766408-4baa-43e9-9a1d-9b0af9fffe61\") " Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.917289 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.917307 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.925324 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26766408-4baa-43e9-9a1d-9b0af9fffe61-kube-api-access-sxbj6" (OuterVolumeSpecName: "kube-api-access-sxbj6") pod "26766408-4baa-43e9-9a1d-9b0af9fffe61" (UID: "26766408-4baa-43e9-9a1d-9b0af9fffe61"). InnerVolumeSpecName "kube-api-access-sxbj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:12:45 crc kubenswrapper[4838]: I1207 09:12:45.961943 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26766408-4baa-43e9-9a1d-9b0af9fffe61" (UID: "26766408-4baa-43e9-9a1d-9b0af9fffe61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.018039 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxbj6\" (UniqueName: \"kubernetes.io/projected/26766408-4baa-43e9-9a1d-9b0af9fffe61-kube-api-access-sxbj6\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.018076 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26766408-4baa-43e9-9a1d-9b0af9fffe61-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.495898 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fh5rj" event={"ID":"26766408-4baa-43e9-9a1d-9b0af9fffe61","Type":"ContainerDied","Data":"610d41efb2d6aef6e3b14157aff12e3c060cee9a0e7a5f5cf7133ebab082941e"} Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.495955 4838 scope.go:117] "RemoveContainer" containerID="4f6b7bb02c589b7341245c2556d6dd34da38cb6169f3849ede8b4f89f0aff990" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.495916 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fh5rj" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.497413 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" event={"ID":"c21136ea-de51-48f3-b79f-493f2d88ece5","Type":"ContainerStarted","Data":"eac00f72e3f626018a3f843df1c577422ac1034c8d16e9d4fdead6002044dcbf"} Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.497440 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" event={"ID":"c21136ea-de51-48f3-b79f-493f2d88ece5","Type":"ContainerStarted","Data":"ba99b834068889a0fbe98b2b0ea88c5e49a5252451e27037b5b133f543e32d64"} Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.497795 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.499054 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" event={"ID":"3854bccd-8788-4cce-8a17-8d06939f172b","Type":"ContainerDied","Data":"f46b97aba20e3c583512c2d9947eee42e7e28f3fb06e661f42ec989cd975f48b"} Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.499154 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll54w" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.503261 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kn79" event={"ID":"be195d3f-68b9-47a0-b726-6e7736450daa","Type":"ContainerDied","Data":"f87889610e05f6898177930b2b5a50808dde796879f494e4b4f5bbb692541bdc"} Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.503330 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kn79" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.512375 4838 scope.go:117] "RemoveContainer" containerID="39f5cf6816c26f9f4f37667af93943767937025255bc60e93401a6835748aaa8" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.512394 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.520968 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d58v2" event={"ID":"b3ab327e-67a8-4c75-809c-9ecb0d4a9be3","Type":"ContainerDied","Data":"922047045a010712980fef48f7b6f2cfcc6b50db30805cdb06e3b52958d2284f"} Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.521015 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d58v2" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.540682 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-69xn8" podStartSLOduration=2.540661836 podStartE2EDuration="2.540661836s" podCreationTimestamp="2025-12-07 09:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:12:46.536252898 +0000 UTC m=+383.243571915" watchObservedRunningTime="2025-12-07 09:12:46.540661836 +0000 UTC m=+383.247980853" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.549946 4838 scope.go:117] "RemoveContainer" containerID="7ad045a7ac4ab70e3a24b2e9b6275ac18c08fa8d3693cdce4771e00f95d54ac7" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.575376 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kn79"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.578004 4838 scope.go:117] "RemoveContainer" containerID="3ec4f3c88edacf90710520d8ed393bdf2fcaa4a691fc0048e77f68fb9a2a37d3" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.582894 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kn79"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.591735 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll54w"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.600577 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll54w"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.605027 4838 scope.go:117] "RemoveContainer" containerID="9f30edaf5cf0fc6cc6b7c4e4bf076e0e2f44b2d156f4fcb792db41d85e062071" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.608267 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fh5rj"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.615664 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fh5rj"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.628889 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d58v2"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.635366 4838 scope.go:117] "RemoveContainer" containerID="93f4bcabc4a7a75ddc83afe4bfcc46f65205f4610423e76a9141df773f40311c" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.642676 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d58v2"] Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.687041 4838 scope.go:117] "RemoveContainer" containerID="4a8649caa7c6d90060b543612b71f6d5c4a689a66e6d42fcdd4cebb4b0caef4d" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.718996 4838 scope.go:117] "RemoveContainer" containerID="cfe1382645341fe7929cd53ccc106774d41ff28adb85f0b5541b9916a9900dc1" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.756183 4838 scope.go:117] "RemoveContainer" containerID="a3308a56e96055b755dfa41811e0e3a85d6a494c166703cc0e62158575698d9d" Dec 07 09:12:46 crc kubenswrapper[4838]: I1207 09:12:46.785864 4838 scope.go:117] "RemoveContainer" containerID="7b4e529a3c267d796322a4c73d179623f7dc29d4bc327ec4d3ab30d155580f9c" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110023 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4pgsb"] Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110426 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110437 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110449 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110455 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110462 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110468 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110475 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110480 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110486 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110491 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110498 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110503 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110512 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110517 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110526 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110531 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110541 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110547 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110556 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110562 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="extract-content" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110569 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110574 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110585 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110592 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110599 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110605 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: E1207 09:12:47.110614 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110620 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="extract-utilities" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110714 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110722 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110730 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110739 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110747 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" containerName="registry-server" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.110956 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" containerName="marketplace-operator" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.111436 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.113740 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.117888 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4pgsb"] Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.237089 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gms9\" (UniqueName: \"kubernetes.io/projected/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-kube-api-access-8gms9\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.237146 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-catalog-content\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.237197 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-utilities\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.338122 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-catalog-content\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.338194 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-utilities\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.338231 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gms9\" (UniqueName: \"kubernetes.io/projected/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-kube-api-access-8gms9\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.338955 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-catalog-content\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.339185 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-utilities\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.355051 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gms9\" (UniqueName: \"kubernetes.io/projected/e7fa6b9c-1255-4cd3-98cb-49600d7f7668-kube-api-access-8gms9\") pod \"redhat-operators-4pgsb\" (UID: \"e7fa6b9c-1255-4cd3-98cb-49600d7f7668\") " pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.426075 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.623620 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26766408-4baa-43e9-9a1d-9b0af9fffe61" path="/var/lib/kubelet/pods/26766408-4baa-43e9-9a1d-9b0af9fffe61/volumes" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.624772 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3854bccd-8788-4cce-8a17-8d06939f172b" path="/var/lib/kubelet/pods/3854bccd-8788-4cce-8a17-8d06939f172b/volumes" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.625467 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ab327e-67a8-4c75-809c-9ecb0d4a9be3" path="/var/lib/kubelet/pods/b3ab327e-67a8-4c75-809c-9ecb0d4a9be3/volumes" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.627093 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be195d3f-68b9-47a0-b726-6e7736450daa" path="/var/lib/kubelet/pods/be195d3f-68b9-47a0-b726-6e7736450daa/volumes" Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.628067 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3208dc2-6618-440a-816c-cab7c4fc394c" path="/var/lib/kubelet/pods/e3208dc2-6618-440a-816c-cab7c4fc394c/volumes" Dec 07 09:12:47 crc kubenswrapper[4838]: W1207 09:12:47.878345 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7fa6b9c_1255_4cd3_98cb_49600d7f7668.slice/crio-1d248ff007086b3024289971997a66469f882a94c820b704dc01e9663e7b5037 WatchSource:0}: Error finding container 1d248ff007086b3024289971997a66469f882a94c820b704dc01e9663e7b5037: Status 404 returned error can't find the container with id 1d248ff007086b3024289971997a66469f882a94c820b704dc01e9663e7b5037 Dec 07 09:12:47 crc kubenswrapper[4838]: I1207 09:12:47.880854 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4pgsb"] Dec 07 09:12:48 crc kubenswrapper[4838]: I1207 09:12:48.564188 4838 generic.go:334] "Generic (PLEG): container finished" podID="e7fa6b9c-1255-4cd3-98cb-49600d7f7668" containerID="789420e7635b4ee7fd8141bdbb973ac8ab466ba17331a89e6283413949647caf" exitCode=0 Dec 07 09:12:48 crc kubenswrapper[4838]: I1207 09:12:48.565800 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pgsb" event={"ID":"e7fa6b9c-1255-4cd3-98cb-49600d7f7668","Type":"ContainerDied","Data":"789420e7635b4ee7fd8141bdbb973ac8ab466ba17331a89e6283413949647caf"} Dec 07 09:12:48 crc kubenswrapper[4838]: I1207 09:12:48.565844 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pgsb" event={"ID":"e7fa6b9c-1255-4cd3-98cb-49600d7f7668","Type":"ContainerStarted","Data":"1d248ff007086b3024289971997a66469f882a94c820b704dc01e9663e7b5037"} Dec 07 09:12:48 crc kubenswrapper[4838]: I1207 09:12:48.923638 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qjfhk"] Dec 07 09:12:48 crc kubenswrapper[4838]: I1207 09:12:48.931791 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qjfhk"] Dec 07 09:12:48 crc kubenswrapper[4838]: I1207 09:12:48.931800 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:48 crc kubenswrapper[4838]: I1207 09:12:48.934456 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.077695 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5989bb7b-629b-47c9-82ae-4e263e313eab-catalog-content\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.078457 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwrgf\" (UniqueName: \"kubernetes.io/projected/5989bb7b-629b-47c9-82ae-4e263e313eab-kube-api-access-hwrgf\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.078540 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5989bb7b-629b-47c9-82ae-4e263e313eab-utilities\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.179842 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5989bb7b-629b-47c9-82ae-4e263e313eab-catalog-content\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.180210 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5989bb7b-629b-47c9-82ae-4e263e313eab-catalog-content\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.180279 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwrgf\" (UniqueName: \"kubernetes.io/projected/5989bb7b-629b-47c9-82ae-4e263e313eab-kube-api-access-hwrgf\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.180299 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5989bb7b-629b-47c9-82ae-4e263e313eab-utilities\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.180541 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5989bb7b-629b-47c9-82ae-4e263e313eab-utilities\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.203121 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwrgf\" (UniqueName: \"kubernetes.io/projected/5989bb7b-629b-47c9-82ae-4e263e313eab-kube-api-access-hwrgf\") pod \"certified-operators-qjfhk\" (UID: \"5989bb7b-629b-47c9-82ae-4e263e313eab\") " pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.264800 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.515687 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7gwg6"] Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.517175 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.519436 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.528640 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gwg6"] Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.578466 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pgsb" event={"ID":"e7fa6b9c-1255-4cd3-98cb-49600d7f7668","Type":"ContainerStarted","Data":"e8d32bc0fe093a1b3dd3e00c64e7994fdc1992e22457cdb042627ae67dde1b3c"} Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.585232 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f62b2408-ca4c-491d-bd9d-335cdd5a5843-catalog-content\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.585356 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq6dw\" (UniqueName: \"kubernetes.io/projected/f62b2408-ca4c-491d-bd9d-335cdd5a5843-kube-api-access-rq6dw\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.585412 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f62b2408-ca4c-491d-bd9d-335cdd5a5843-utilities\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.675186 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qjfhk"] Dec 07 09:12:49 crc kubenswrapper[4838]: W1207 09:12:49.679648 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5989bb7b_629b_47c9_82ae_4e263e313eab.slice/crio-120e9188dc8acc9cab18a153248782b1fe49a5da8792a3189897400be9e39924 WatchSource:0}: Error finding container 120e9188dc8acc9cab18a153248782b1fe49a5da8792a3189897400be9e39924: Status 404 returned error can't find the container with id 120e9188dc8acc9cab18a153248782b1fe49a5da8792a3189897400be9e39924 Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.686384 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq6dw\" (UniqueName: \"kubernetes.io/projected/f62b2408-ca4c-491d-bd9d-335cdd5a5843-kube-api-access-rq6dw\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.686454 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f62b2408-ca4c-491d-bd9d-335cdd5a5843-utilities\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.686513 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f62b2408-ca4c-491d-bd9d-335cdd5a5843-catalog-content\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.687312 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f62b2408-ca4c-491d-bd9d-335cdd5a5843-utilities\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.688085 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f62b2408-ca4c-491d-bd9d-335cdd5a5843-catalog-content\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.708399 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq6dw\" (UniqueName: \"kubernetes.io/projected/f62b2408-ca4c-491d-bd9d-335cdd5a5843-kube-api-access-rq6dw\") pod \"community-operators-7gwg6\" (UID: \"f62b2408-ca4c-491d-bd9d-335cdd5a5843\") " pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:49 crc kubenswrapper[4838]: I1207 09:12:49.833424 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.208737 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gwg6"] Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.584704 4838 generic.go:334] "Generic (PLEG): container finished" podID="e7fa6b9c-1255-4cd3-98cb-49600d7f7668" containerID="e8d32bc0fe093a1b3dd3e00c64e7994fdc1992e22457cdb042627ae67dde1b3c" exitCode=0 Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.584758 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pgsb" event={"ID":"e7fa6b9c-1255-4cd3-98cb-49600d7f7668","Type":"ContainerDied","Data":"e8d32bc0fe093a1b3dd3e00c64e7994fdc1992e22457cdb042627ae67dde1b3c"} Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.587035 4838 generic.go:334] "Generic (PLEG): container finished" podID="f62b2408-ca4c-491d-bd9d-335cdd5a5843" containerID="b5316c94ece9b44381f1cef83fdd0ff9aec6efb5c99e4f87c28fa7a5431074bf" exitCode=0 Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.587087 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gwg6" event={"ID":"f62b2408-ca4c-491d-bd9d-335cdd5a5843","Type":"ContainerDied","Data":"b5316c94ece9b44381f1cef83fdd0ff9aec6efb5c99e4f87c28fa7a5431074bf"} Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.587109 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gwg6" event={"ID":"f62b2408-ca4c-491d-bd9d-335cdd5a5843","Type":"ContainerStarted","Data":"25298c756aeaa00e9e0efde2b5d3bb9e5017b8e623595e2ca033fa686dc471a6"} Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.590444 4838 generic.go:334] "Generic (PLEG): container finished" podID="5989bb7b-629b-47c9-82ae-4e263e313eab" containerID="81eb3955f29f0272e127196bd947b01d0dc4dce8355c880b8bb62ef7785dc899" exitCode=0 Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.590810 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjfhk" event={"ID":"5989bb7b-629b-47c9-82ae-4e263e313eab","Type":"ContainerDied","Data":"81eb3955f29f0272e127196bd947b01d0dc4dce8355c880b8bb62ef7785dc899"} Dec 07 09:12:50 crc kubenswrapper[4838]: I1207 09:12:50.590859 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjfhk" event={"ID":"5989bb7b-629b-47c9-82ae-4e263e313eab","Type":"ContainerStarted","Data":"120e9188dc8acc9cab18a153248782b1fe49a5da8792a3189897400be9e39924"} Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.313937 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwqp"] Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.314847 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.317146 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.332809 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwqp"] Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.503271 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-utilities\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.503372 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqw5m\" (UniqueName: \"kubernetes.io/projected/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-kube-api-access-lqw5m\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.503406 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-catalog-content\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.597582 4838 generic.go:334] "Generic (PLEG): container finished" podID="5989bb7b-629b-47c9-82ae-4e263e313eab" containerID="a50edfb24a3dcb2bc78b57f25990513d9230739ae12aad845d335b734877fba7" exitCode=0 Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.597663 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjfhk" event={"ID":"5989bb7b-629b-47c9-82ae-4e263e313eab","Type":"ContainerDied","Data":"a50edfb24a3dcb2bc78b57f25990513d9230739ae12aad845d335b734877fba7"} Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.605802 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqw5m\" (UniqueName: \"kubernetes.io/projected/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-kube-api-access-lqw5m\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.605874 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-catalog-content\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.605958 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-utilities\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.606404 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pgsb" event={"ID":"e7fa6b9c-1255-4cd3-98cb-49600d7f7668","Type":"ContainerStarted","Data":"03dd8f378dd03571e973decbdd396d692c0e2a4465ae8417012dbfbe6a09c520"} Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.606447 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-utilities\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.606989 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-catalog-content\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.608064 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gwg6" event={"ID":"f62b2408-ca4c-491d-bd9d-335cdd5a5843","Type":"ContainerStarted","Data":"470ef2b2348c025bf382ad59838d68ae3a3c9b62eab04d74b28a9075ca1fc750"} Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.634222 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqw5m\" (UniqueName: \"kubernetes.io/projected/6ba85040-ab6e-46d8-b6ce-ea49e3a3311c-kube-api-access-lqw5m\") pod \"redhat-marketplace-dwwqp\" (UID: \"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c\") " pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.713420 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4pgsb" podStartSLOduration=2.2430917 podStartE2EDuration="4.713398549s" podCreationTimestamp="2025-12-07 09:12:47 +0000 UTC" firstStartedPulling="2025-12-07 09:12:48.566094424 +0000 UTC m=+385.273413441" lastFinishedPulling="2025-12-07 09:12:51.036401273 +0000 UTC m=+387.743720290" observedRunningTime="2025-12-07 09:12:51.708429164 +0000 UTC m=+388.415748201" watchObservedRunningTime="2025-12-07 09:12:51.713398549 +0000 UTC m=+388.420717566" Dec 07 09:12:51 crc kubenswrapper[4838]: I1207 09:12:51.927340 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.322462 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwqp"] Dec 07 09:12:52 crc kubenswrapper[4838]: W1207 09:12:52.331384 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ba85040_ab6e_46d8_b6ce_ea49e3a3311c.slice/crio-bc4e90363b107f39a3cdfece4f528b125576d2300ea0c1b4951f31d0f1236077 WatchSource:0}: Error finding container bc4e90363b107f39a3cdfece4f528b125576d2300ea0c1b4951f31d0f1236077: Status 404 returned error can't find the container with id bc4e90363b107f39a3cdfece4f528b125576d2300ea0c1b4951f31d0f1236077 Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.614979 4838 generic.go:334] "Generic (PLEG): container finished" podID="f62b2408-ca4c-491d-bd9d-335cdd5a5843" containerID="470ef2b2348c025bf382ad59838d68ae3a3c9b62eab04d74b28a9075ca1fc750" exitCode=0 Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.615190 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gwg6" event={"ID":"f62b2408-ca4c-491d-bd9d-335cdd5a5843","Type":"ContainerDied","Data":"470ef2b2348c025bf382ad59838d68ae3a3c9b62eab04d74b28a9075ca1fc750"} Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.617377 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjfhk" event={"ID":"5989bb7b-629b-47c9-82ae-4e263e313eab","Type":"ContainerStarted","Data":"a2ca9bdb31d3f54958a5ce27053f6b6945ce71b0cad0ccc8305ff41107651d5c"} Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.619451 4838 generic.go:334] "Generic (PLEG): container finished" podID="6ba85040-ab6e-46d8-b6ce-ea49e3a3311c" containerID="c62a4cc9b928f082351c448a8729b655f235fde530575862edb62f8984097d44" exitCode=0 Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.619854 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwqp" event={"ID":"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c","Type":"ContainerDied","Data":"c62a4cc9b928f082351c448a8729b655f235fde530575862edb62f8984097d44"} Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.619876 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwqp" event={"ID":"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c","Type":"ContainerStarted","Data":"bc4e90363b107f39a3cdfece4f528b125576d2300ea0c1b4951f31d0f1236077"} Dec 07 09:12:52 crc kubenswrapper[4838]: I1207 09:12:52.655369 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qjfhk" podStartSLOduration=3.240125812 podStartE2EDuration="4.65534948s" podCreationTimestamp="2025-12-07 09:12:48 +0000 UTC" firstStartedPulling="2025-12-07 09:12:50.592155401 +0000 UTC m=+387.299474418" lastFinishedPulling="2025-12-07 09:12:52.007379049 +0000 UTC m=+388.714698086" observedRunningTime="2025-12-07 09:12:52.652860708 +0000 UTC m=+389.360179735" watchObservedRunningTime="2025-12-07 09:12:52.65534948 +0000 UTC m=+389.362668497" Dec 07 09:12:53 crc kubenswrapper[4838]: I1207 09:12:53.626065 4838 generic.go:334] "Generic (PLEG): container finished" podID="6ba85040-ab6e-46d8-b6ce-ea49e3a3311c" containerID="9d5ada1ab3476bd6af4b7b6962db1cbb76b849be613d1f2e1dce5e57f59317ca" exitCode=0 Dec 07 09:12:53 crc kubenswrapper[4838]: I1207 09:12:53.626147 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwqp" event={"ID":"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c","Type":"ContainerDied","Data":"9d5ada1ab3476bd6af4b7b6962db1cbb76b849be613d1f2e1dce5e57f59317ca"} Dec 07 09:12:53 crc kubenswrapper[4838]: I1207 09:12:53.630964 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gwg6" event={"ID":"f62b2408-ca4c-491d-bd9d-335cdd5a5843","Type":"ContainerStarted","Data":"2c42f5d714eef3fc0d574c44e50cfcad095cdf4317f35c67eec12fe8f8d889a2"} Dec 07 09:12:53 crc kubenswrapper[4838]: I1207 09:12:53.694893 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7gwg6" podStartSLOduration=2.298483433 podStartE2EDuration="4.694871556s" podCreationTimestamp="2025-12-07 09:12:49 +0000 UTC" firstStartedPulling="2025-12-07 09:12:50.588616648 +0000 UTC m=+387.295935665" lastFinishedPulling="2025-12-07 09:12:52.985004771 +0000 UTC m=+389.692323788" observedRunningTime="2025-12-07 09:12:53.691264571 +0000 UTC m=+390.398583588" watchObservedRunningTime="2025-12-07 09:12:53.694871556 +0000 UTC m=+390.402190583" Dec 07 09:12:54 crc kubenswrapper[4838]: I1207 09:12:54.135829 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hk46c" Dec 07 09:12:54 crc kubenswrapper[4838]: I1207 09:12:54.226944 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-66ndx"] Dec 07 09:12:54 crc kubenswrapper[4838]: I1207 09:12:54.493521 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:12:54 crc kubenswrapper[4838]: I1207 09:12:54.493872 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:12:54 crc kubenswrapper[4838]: I1207 09:12:54.493918 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:12:54 crc kubenswrapper[4838]: I1207 09:12:54.494460 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5071e5fc8badded20c88d60e74e442b4c8646f5320a60585369874896a169785"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:12:54 crc kubenswrapper[4838]: I1207 09:12:54.494513 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://5071e5fc8badded20c88d60e74e442b4c8646f5320a60585369874896a169785" gracePeriod=600 Dec 07 09:12:55 crc kubenswrapper[4838]: I1207 09:12:55.643389 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="5071e5fc8badded20c88d60e74e442b4c8646f5320a60585369874896a169785" exitCode=0 Dec 07 09:12:55 crc kubenswrapper[4838]: I1207 09:12:55.643949 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"5071e5fc8badded20c88d60e74e442b4c8646f5320a60585369874896a169785"} Dec 07 09:12:55 crc kubenswrapper[4838]: I1207 09:12:55.643977 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"f5608e3e5fc8b2067546b7fff7ad2e9ceab4038aa8bfb537b0261a0f743498fd"} Dec 07 09:12:55 crc kubenswrapper[4838]: I1207 09:12:55.643993 4838 scope.go:117] "RemoveContainer" containerID="c92d20f9dd3db97a3c66507468d8f6f0a9c71c324c7a081c3177b5ea727dfd67" Dec 07 09:12:55 crc kubenswrapper[4838]: I1207 09:12:55.648620 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwqp" event={"ID":"6ba85040-ab6e-46d8-b6ce-ea49e3a3311c","Type":"ContainerStarted","Data":"211fd42d5c0396d034ca8fe0b91a32c47b5d48ec25a09811dc2ce2b108aa1210"} Dec 07 09:12:55 crc kubenswrapper[4838]: I1207 09:12:55.683616 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dwwqp" podStartSLOduration=3.285108116 podStartE2EDuration="4.683598147s" podCreationTimestamp="2025-12-07 09:12:51 +0000 UTC" firstStartedPulling="2025-12-07 09:12:52.620968318 +0000 UTC m=+389.328287335" lastFinishedPulling="2025-12-07 09:12:54.019458359 +0000 UTC m=+390.726777366" observedRunningTime="2025-12-07 09:12:55.679035197 +0000 UTC m=+392.386354214" watchObservedRunningTime="2025-12-07 09:12:55.683598147 +0000 UTC m=+392.390917164" Dec 07 09:12:57 crc kubenswrapper[4838]: I1207 09:12:57.426543 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:57 crc kubenswrapper[4838]: I1207 09:12:57.427052 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:57 crc kubenswrapper[4838]: I1207 09:12:57.478357 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:57 crc kubenswrapper[4838]: I1207 09:12:57.709847 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4pgsb" Dec 07 09:12:59 crc kubenswrapper[4838]: I1207 09:12:59.265618 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:59 crc kubenswrapper[4838]: I1207 09:12:59.265923 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:59 crc kubenswrapper[4838]: I1207 09:12:59.309077 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:59 crc kubenswrapper[4838]: I1207 09:12:59.717077 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qjfhk" Dec 07 09:12:59 crc kubenswrapper[4838]: I1207 09:12:59.834531 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:59 crc kubenswrapper[4838]: I1207 09:12:59.834586 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:12:59 crc kubenswrapper[4838]: I1207 09:12:59.883451 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:13:00 crc kubenswrapper[4838]: I1207 09:13:00.738713 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7gwg6" Dec 07 09:13:01 crc kubenswrapper[4838]: I1207 09:13:01.928834 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:13:01 crc kubenswrapper[4838]: I1207 09:13:01.928884 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:13:01 crc kubenswrapper[4838]: I1207 09:13:01.967064 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:13:02 crc kubenswrapper[4838]: I1207 09:13:02.763125 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dwwqp" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.274398 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" podUID="725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" containerName="registry" containerID="cri-o://7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726" gracePeriod=30 Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.706400 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815306 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-tls\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815358 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-bound-sa-token\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815513 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815549 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-installation-pull-secrets\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815602 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-certificates\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815624 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-ca-trust-extracted\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815675 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-trusted-ca\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.815696 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvhmh\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-kube-api-access-bvhmh\") pod \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\" (UID: \"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb\") " Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.816456 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.817438 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.822378 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-kube-api-access-bvhmh" (OuterVolumeSpecName: "kube-api-access-bvhmh") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "kube-api-access-bvhmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.822610 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.823246 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.824590 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.831589 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.837033 4838 generic.go:334] "Generic (PLEG): container finished" podID="725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" containerID="7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726" exitCode=0 Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.837175 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" event={"ID":"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb","Type":"ContainerDied","Data":"7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726"} Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.837271 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" event={"ID":"725baa02-2d4f-4ef1-aa67-3d26aff3f5bb","Type":"ContainerDied","Data":"df54c7d011c4f3ad9f55ee68e58eb45339431009519910f3f32d3a14e5a20818"} Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.837499 4838 scope.go:117] "RemoveContainer" containerID="7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.837685 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-66ndx" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.845536 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" (UID: "725baa02-2d4f-4ef1-aa67-3d26aff3f5bb"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.871105 4838 scope.go:117] "RemoveContainer" containerID="7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726" Dec 07 09:13:19 crc kubenswrapper[4838]: E1207 09:13:19.872251 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726\": container with ID starting with 7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726 not found: ID does not exist" containerID="7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.872295 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726"} err="failed to get container status \"7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726\": rpc error: code = NotFound desc = could not find container \"7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726\": container with ID starting with 7857ff8b2805ecd10eb9a2ecbfb3e07f4d43b913f6fb7949f89478ae929cc726 not found: ID does not exist" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.917110 4838 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.917338 4838 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.917431 4838 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.917498 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.917560 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvhmh\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-kube-api-access-bvhmh\") on node \"crc\" DevicePath \"\"" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.917614 4838 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:13:19 crc kubenswrapper[4838]: I1207 09:13:19.917667 4838 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 09:13:20 crc kubenswrapper[4838]: I1207 09:13:20.194024 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-66ndx"] Dec 07 09:13:20 crc kubenswrapper[4838]: I1207 09:13:20.198614 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-66ndx"] Dec 07 09:13:21 crc kubenswrapper[4838]: I1207 09:13:21.627598 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" path="/var/lib/kubelet/pods/725baa02-2d4f-4ef1-aa67-3d26aff3f5bb/volumes" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.184010 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q"] Dec 07 09:15:00 crc kubenswrapper[4838]: E1207 09:15:00.184954 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" containerName="registry" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.184978 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" containerName="registry" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.185149 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="725baa02-2d4f-4ef1-aa67-3d26aff3f5bb" containerName="registry" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.185682 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.188537 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.192537 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.193352 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q"] Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.383635 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd29j\" (UniqueName: \"kubernetes.io/projected/9083d71b-2831-4a45-80b2-baa769ed181f-kube-api-access-jd29j\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.383719 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9083d71b-2831-4a45-80b2-baa769ed181f-config-volume\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.383866 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9083d71b-2831-4a45-80b2-baa769ed181f-secret-volume\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.484436 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9083d71b-2831-4a45-80b2-baa769ed181f-config-volume\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.484489 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9083d71b-2831-4a45-80b2-baa769ed181f-secret-volume\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.484529 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd29j\" (UniqueName: \"kubernetes.io/projected/9083d71b-2831-4a45-80b2-baa769ed181f-kube-api-access-jd29j\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.486718 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9083d71b-2831-4a45-80b2-baa769ed181f-config-volume\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.493111 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9083d71b-2831-4a45-80b2-baa769ed181f-secret-volume\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.505253 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd29j\" (UniqueName: \"kubernetes.io/projected/9083d71b-2831-4a45-80b2-baa769ed181f-kube-api-access-jd29j\") pod \"collect-profiles-29418315-4ph5q\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.523830 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:00 crc kubenswrapper[4838]: I1207 09:15:00.707415 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q"] Dec 07 09:15:01 crc kubenswrapper[4838]: I1207 09:15:01.502525 4838 generic.go:334] "Generic (PLEG): container finished" podID="9083d71b-2831-4a45-80b2-baa769ed181f" containerID="617037b1b5c9e73c95d2534390edb323f8fad7af92a6df7183339bf2463858e9" exitCode=0 Dec 07 09:15:01 crc kubenswrapper[4838]: I1207 09:15:01.502855 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" event={"ID":"9083d71b-2831-4a45-80b2-baa769ed181f","Type":"ContainerDied","Data":"617037b1b5c9e73c95d2534390edb323f8fad7af92a6df7183339bf2463858e9"} Dec 07 09:15:01 crc kubenswrapper[4838]: I1207 09:15:01.502890 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" event={"ID":"9083d71b-2831-4a45-80b2-baa769ed181f","Type":"ContainerStarted","Data":"527aadabc7068c186c3bbb36358210cbabbe5a3eaa836ad5217e56c49c04aec9"} Dec 07 09:15:02 crc kubenswrapper[4838]: I1207 09:15:02.754921 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:02 crc kubenswrapper[4838]: I1207 09:15:02.925521 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9083d71b-2831-4a45-80b2-baa769ed181f-config-volume\") pod \"9083d71b-2831-4a45-80b2-baa769ed181f\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " Dec 07 09:15:02 crc kubenswrapper[4838]: I1207 09:15:02.925644 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9083d71b-2831-4a45-80b2-baa769ed181f-secret-volume\") pod \"9083d71b-2831-4a45-80b2-baa769ed181f\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " Dec 07 09:15:02 crc kubenswrapper[4838]: I1207 09:15:02.925700 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd29j\" (UniqueName: \"kubernetes.io/projected/9083d71b-2831-4a45-80b2-baa769ed181f-kube-api-access-jd29j\") pod \"9083d71b-2831-4a45-80b2-baa769ed181f\" (UID: \"9083d71b-2831-4a45-80b2-baa769ed181f\") " Dec 07 09:15:02 crc kubenswrapper[4838]: I1207 09:15:02.926548 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9083d71b-2831-4a45-80b2-baa769ed181f-config-volume" (OuterVolumeSpecName: "config-volume") pod "9083d71b-2831-4a45-80b2-baa769ed181f" (UID: "9083d71b-2831-4a45-80b2-baa769ed181f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:15:02 crc kubenswrapper[4838]: I1207 09:15:02.931423 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9083d71b-2831-4a45-80b2-baa769ed181f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9083d71b-2831-4a45-80b2-baa769ed181f" (UID: "9083d71b-2831-4a45-80b2-baa769ed181f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:15:02 crc kubenswrapper[4838]: I1207 09:15:02.931953 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9083d71b-2831-4a45-80b2-baa769ed181f-kube-api-access-jd29j" (OuterVolumeSpecName: "kube-api-access-jd29j") pod "9083d71b-2831-4a45-80b2-baa769ed181f" (UID: "9083d71b-2831-4a45-80b2-baa769ed181f"). InnerVolumeSpecName "kube-api-access-jd29j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:15:03 crc kubenswrapper[4838]: I1207 09:15:03.026896 4838 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9083d71b-2831-4a45-80b2-baa769ed181f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:15:03 crc kubenswrapper[4838]: I1207 09:15:03.027070 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd29j\" (UniqueName: \"kubernetes.io/projected/9083d71b-2831-4a45-80b2-baa769ed181f-kube-api-access-jd29j\") on node \"crc\" DevicePath \"\"" Dec 07 09:15:03 crc kubenswrapper[4838]: I1207 09:15:03.027121 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9083d71b-2831-4a45-80b2-baa769ed181f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:15:03 crc kubenswrapper[4838]: I1207 09:15:03.518989 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" event={"ID":"9083d71b-2831-4a45-80b2-baa769ed181f","Type":"ContainerDied","Data":"527aadabc7068c186c3bbb36358210cbabbe5a3eaa836ad5217e56c49c04aec9"} Dec 07 09:15:03 crc kubenswrapper[4838]: I1207 09:15:03.519373 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="527aadabc7068c186c3bbb36358210cbabbe5a3eaa836ad5217e56c49c04aec9" Dec 07 09:15:03 crc kubenswrapper[4838]: I1207 09:15:03.519074 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q" Dec 07 09:15:24 crc kubenswrapper[4838]: I1207 09:15:24.494045 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:15:24 crc kubenswrapper[4838]: I1207 09:15:24.494755 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:15:54 crc kubenswrapper[4838]: I1207 09:15:54.493505 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:15:54 crc kubenswrapper[4838]: I1207 09:15:54.494125 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:16:24 crc kubenswrapper[4838]: I1207 09:16:24.493922 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:16:24 crc kubenswrapper[4838]: I1207 09:16:24.494645 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:16:24 crc kubenswrapper[4838]: I1207 09:16:24.494726 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:16:24 crc kubenswrapper[4838]: I1207 09:16:24.500269 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5608e3e5fc8b2067546b7fff7ad2e9ceab4038aa8bfb537b0261a0f743498fd"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:16:24 crc kubenswrapper[4838]: I1207 09:16:24.500439 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://f5608e3e5fc8b2067546b7fff7ad2e9ceab4038aa8bfb537b0261a0f743498fd" gracePeriod=600 Dec 07 09:16:25 crc kubenswrapper[4838]: I1207 09:16:25.050735 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="f5608e3e5fc8b2067546b7fff7ad2e9ceab4038aa8bfb537b0261a0f743498fd" exitCode=0 Dec 07 09:16:25 crc kubenswrapper[4838]: I1207 09:16:25.050768 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"f5608e3e5fc8b2067546b7fff7ad2e9ceab4038aa8bfb537b0261a0f743498fd"} Dec 07 09:16:25 crc kubenswrapper[4838]: I1207 09:16:25.051139 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"4e006964b42caddec4f66d1cf7932894d0547718b0bf611546057b7159422840"} Dec 07 09:16:25 crc kubenswrapper[4838]: I1207 09:16:25.051167 4838 scope.go:117] "RemoveContainer" containerID="5071e5fc8badded20c88d60e74e442b4c8646f5320a60585369874896a169785" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.131602 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6knh9"] Dec 07 09:17:53 crc kubenswrapper[4838]: E1207 09:17:53.132435 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9083d71b-2831-4a45-80b2-baa769ed181f" containerName="collect-profiles" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.132453 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9083d71b-2831-4a45-80b2-baa769ed181f" containerName="collect-profiles" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.132573 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9083d71b-2831-4a45-80b2-baa769ed181f" containerName="collect-profiles" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.133076 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.139367 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6knh9"] Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.140904 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.141112 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.141226 4838 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7lhks" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.153036 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2h6mq"] Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.153640 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2h6mq" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.155920 4838 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-mndhb" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.179740 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2h6mq"] Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.185026 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-87mkv"] Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.185689 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.187609 4838 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-ljd89" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.194532 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-87mkv"] Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.240438 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7xd9\" (UniqueName: \"kubernetes.io/projected/72c34e1b-8df3-4622-9cdb-b44adf0e45bf-kube-api-access-j7xd9\") pod \"cert-manager-webhook-5655c58dd6-87mkv\" (UID: \"72c34e1b-8df3-4622-9cdb-b44adf0e45bf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.240509 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7mgb\" (UniqueName: \"kubernetes.io/projected/c0fb9b54-1b0a-4cab-8bad-669c11ea8a73-kube-api-access-m7mgb\") pod \"cert-manager-cainjector-7f985d654d-6knh9\" (UID: \"c0fb9b54-1b0a-4cab-8bad-669c11ea8a73\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.240530 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4frwc\" (UniqueName: \"kubernetes.io/projected/d7e66c54-9769-4287-8b1c-f1d1a56b08a1-kube-api-access-4frwc\") pod \"cert-manager-5b446d88c5-2h6mq\" (UID: \"d7e66c54-9769-4287-8b1c-f1d1a56b08a1\") " pod="cert-manager/cert-manager-5b446d88c5-2h6mq" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.341718 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7xd9\" (UniqueName: \"kubernetes.io/projected/72c34e1b-8df3-4622-9cdb-b44adf0e45bf-kube-api-access-j7xd9\") pod \"cert-manager-webhook-5655c58dd6-87mkv\" (UID: \"72c34e1b-8df3-4622-9cdb-b44adf0e45bf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.341801 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7mgb\" (UniqueName: \"kubernetes.io/projected/c0fb9b54-1b0a-4cab-8bad-669c11ea8a73-kube-api-access-m7mgb\") pod \"cert-manager-cainjector-7f985d654d-6knh9\" (UID: \"c0fb9b54-1b0a-4cab-8bad-669c11ea8a73\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.341847 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4frwc\" (UniqueName: \"kubernetes.io/projected/d7e66c54-9769-4287-8b1c-f1d1a56b08a1-kube-api-access-4frwc\") pod \"cert-manager-5b446d88c5-2h6mq\" (UID: \"d7e66c54-9769-4287-8b1c-f1d1a56b08a1\") " pod="cert-manager/cert-manager-5b446d88c5-2h6mq" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.361600 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7mgb\" (UniqueName: \"kubernetes.io/projected/c0fb9b54-1b0a-4cab-8bad-669c11ea8a73-kube-api-access-m7mgb\") pod \"cert-manager-cainjector-7f985d654d-6knh9\" (UID: \"c0fb9b54-1b0a-4cab-8bad-669c11ea8a73\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.361879 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7xd9\" (UniqueName: \"kubernetes.io/projected/72c34e1b-8df3-4622-9cdb-b44adf0e45bf-kube-api-access-j7xd9\") pod \"cert-manager-webhook-5655c58dd6-87mkv\" (UID: \"72c34e1b-8df3-4622-9cdb-b44adf0e45bf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.362754 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4frwc\" (UniqueName: \"kubernetes.io/projected/d7e66c54-9769-4287-8b1c-f1d1a56b08a1-kube-api-access-4frwc\") pod \"cert-manager-5b446d88c5-2h6mq\" (UID: \"d7e66c54-9769-4287-8b1c-f1d1a56b08a1\") " pod="cert-manager/cert-manager-5b446d88c5-2h6mq" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.457733 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.468510 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2h6mq" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.503585 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.772318 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-87mkv"] Dec 07 09:17:53 crc kubenswrapper[4838]: W1207 09:17:53.775678 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72c34e1b_8df3_4622_9cdb_b44adf0e45bf.slice/crio-9699e69c85affacef4c739401b7b618bdd7b67cf4a217e7a005c1399e2f36219 WatchSource:0}: Error finding container 9699e69c85affacef4c739401b7b618bdd7b67cf4a217e7a005c1399e2f36219: Status 404 returned error can't find the container with id 9699e69c85affacef4c739401b7b618bdd7b67cf4a217e7a005c1399e2f36219 Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.777361 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.927603 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2h6mq"] Dec 07 09:17:53 crc kubenswrapper[4838]: I1207 09:17:53.932011 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6knh9"] Dec 07 09:17:53 crc kubenswrapper[4838]: W1207 09:17:53.938223 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0fb9b54_1b0a_4cab_8bad_669c11ea8a73.slice/crio-30b97072911f494722017fd4c6403520195983472a917533b040c4a82f10d130 WatchSource:0}: Error finding container 30b97072911f494722017fd4c6403520195983472a917533b040c4a82f10d130: Status 404 returned error can't find the container with id 30b97072911f494722017fd4c6403520195983472a917533b040c4a82f10d130 Dec 07 09:17:53 crc kubenswrapper[4838]: W1207 09:17:53.940803 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7e66c54_9769_4287_8b1c_f1d1a56b08a1.slice/crio-2632f3d6d0ab42496326e007257a8197408d95c7d1e89a7ac06d25a245592e80 WatchSource:0}: Error finding container 2632f3d6d0ab42496326e007257a8197408d95c7d1e89a7ac06d25a245592e80: Status 404 returned error can't find the container with id 2632f3d6d0ab42496326e007257a8197408d95c7d1e89a7ac06d25a245592e80 Dec 07 09:17:54 crc kubenswrapper[4838]: I1207 09:17:54.640102 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2h6mq" event={"ID":"d7e66c54-9769-4287-8b1c-f1d1a56b08a1","Type":"ContainerStarted","Data":"2632f3d6d0ab42496326e007257a8197408d95c7d1e89a7ac06d25a245592e80"} Dec 07 09:17:54 crc kubenswrapper[4838]: I1207 09:17:54.650573 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" event={"ID":"c0fb9b54-1b0a-4cab-8bad-669c11ea8a73","Type":"ContainerStarted","Data":"30b97072911f494722017fd4c6403520195983472a917533b040c4a82f10d130"} Dec 07 09:17:54 crc kubenswrapper[4838]: I1207 09:17:54.652062 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" event={"ID":"72c34e1b-8df3-4622-9cdb-b44adf0e45bf","Type":"ContainerStarted","Data":"9699e69c85affacef4c739401b7b618bdd7b67cf4a217e7a005c1399e2f36219"} Dec 07 09:17:57 crc kubenswrapper[4838]: I1207 09:17:57.676122 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2h6mq" event={"ID":"d7e66c54-9769-4287-8b1c-f1d1a56b08a1","Type":"ContainerStarted","Data":"f9cbec2f32a67f1673f8784b00646c12c7de798f35b9edea7ef8cbf6545e172e"} Dec 07 09:17:57 crc kubenswrapper[4838]: I1207 09:17:57.678296 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" event={"ID":"c0fb9b54-1b0a-4cab-8bad-669c11ea8a73","Type":"ContainerStarted","Data":"15d823e04fa2657fe52efe04a9b1aafc557d524d94cf2dc2b27b1b0cf8bb6918"} Dec 07 09:17:57 crc kubenswrapper[4838]: I1207 09:17:57.682884 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" event={"ID":"72c34e1b-8df3-4622-9cdb-b44adf0e45bf","Type":"ContainerStarted","Data":"d0f65d1550b45f73f0c847ed94ee2076cd95b63ec60503372efa9be1edeaed60"} Dec 07 09:17:57 crc kubenswrapper[4838]: I1207 09:17:57.683085 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" Dec 07 09:17:57 crc kubenswrapper[4838]: I1207 09:17:57.708683 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-2h6mq" podStartSLOduration=2.028922149 podStartE2EDuration="4.7086611s" podCreationTimestamp="2025-12-07 09:17:53 +0000 UTC" firstStartedPulling="2025-12-07 09:17:53.94401569 +0000 UTC m=+690.651334717" lastFinishedPulling="2025-12-07 09:17:56.623754621 +0000 UTC m=+693.331073668" observedRunningTime="2025-12-07 09:17:57.698923735 +0000 UTC m=+694.406242772" watchObservedRunningTime="2025-12-07 09:17:57.7086611 +0000 UTC m=+694.415980127" Dec 07 09:17:57 crc kubenswrapper[4838]: I1207 09:17:57.730131 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-6knh9" podStartSLOduration=2.047526874 podStartE2EDuration="4.730113186s" podCreationTimestamp="2025-12-07 09:17:53 +0000 UTC" firstStartedPulling="2025-12-07 09:17:53.939714578 +0000 UTC m=+690.647033615" lastFinishedPulling="2025-12-07 09:17:56.62230088 +0000 UTC m=+693.329619927" observedRunningTime="2025-12-07 09:17:57.726846724 +0000 UTC m=+694.434165751" watchObservedRunningTime="2025-12-07 09:17:57.730113186 +0000 UTC m=+694.437432223" Dec 07 09:17:57 crc kubenswrapper[4838]: I1207 09:17:57.748946 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" podStartSLOduration=1.922905141 podStartE2EDuration="4.748928348s" podCreationTimestamp="2025-12-07 09:17:53 +0000 UTC" firstStartedPulling="2025-12-07 09:17:53.777156913 +0000 UTC m=+690.484475930" lastFinishedPulling="2025-12-07 09:17:56.6031801 +0000 UTC m=+693.310499137" observedRunningTime="2025-12-07 09:17:57.748146066 +0000 UTC m=+694.455465083" watchObservedRunningTime="2025-12-07 09:17:57.748928348 +0000 UTC m=+694.456247365" Dec 07 09:18:03 crc kubenswrapper[4838]: I1207 09:18:03.507030 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-87mkv" Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.873946 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-brr6g"] Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.875318 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-controller" containerID="cri-o://d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" gracePeriod=30 Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.875435 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-node" containerID="cri-o://8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" gracePeriod=30 Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.875476 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="sbdb" containerID="cri-o://2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" gracePeriod=30 Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.875568 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-acl-logging" containerID="cri-o://6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" gracePeriod=30 Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.875637 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="northd" containerID="cri-o://d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" gracePeriod=30 Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.875692 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" gracePeriod=30 Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.876012 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="nbdb" containerID="cri-o://96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" gracePeriod=30 Dec 07 09:18:15 crc kubenswrapper[4838]: I1207 09:18:15.919534 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" containerID="cri-o://ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" gracePeriod=30 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.268219 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/3.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.270071 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovn-acl-logging/0.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.270906 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovn-controller/0.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.271341 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.325900 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-env-overrides\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.325958 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-ovn\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.325984 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-systemd-units\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326004 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-var-lib-openvswitch\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326027 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-netd\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326053 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-slash\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326075 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-config\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326097 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-kubelet\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326115 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-netns\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326140 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-node-log\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326159 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-systemd\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326183 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66f9850-4704-46f3-9745-bdf80caf8a98-ovn-node-metrics-cert\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326223 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-openvswitch\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326248 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-etc-openvswitch\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326274 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcrtj\" (UniqueName: \"kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326319 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-script-lib\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326340 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-bin\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326357 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-ovn-kubernetes\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326388 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-log-socket\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326407 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a66f9850-4704-46f3-9745-bdf80caf8a98\" (UID: \"a66f9850-4704-46f3-9745-bdf80caf8a98\") " Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.326609 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.327986 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328033 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328084 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328107 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328128 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328150 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-slash" (OuterVolumeSpecName: "host-slash") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328800 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328856 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328881 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.328906 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-node-log" (OuterVolumeSpecName: "node-log") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.329710 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cbtsn"] Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.329952 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.329977 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.329987 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330040 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330085 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330059 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330114 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-log-socket" (OuterVolumeSpecName: "log-socket") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330149 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330553 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330069 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330596 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330603 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330612 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-ovn-metrics" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330619 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-ovn-metrics" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330637 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="sbdb" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330645 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="sbdb" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330661 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kubecfg-setup" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330668 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kubecfg-setup" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330678 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330686 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330696 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-acl-logging" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330704 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-acl-logging" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330715 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330723 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330733 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="nbdb" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330740 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="nbdb" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330751 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-node" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330759 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-node" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.330770 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="northd" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330778 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="northd" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330916 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="northd" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330932 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-ovn-metrics" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330944 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-acl-logging" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330954 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330962 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="kube-rbac-proxy-node" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330973 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330983 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.330994 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.331002 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="nbdb" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.331011 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovn-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.331020 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="sbdb" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.331138 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.331148 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.331270 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerName="ovnkube-controller" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.333257 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.339278 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66f9850-4704-46f3-9745-bdf80caf8a98-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.339387 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj" (OuterVolumeSpecName: "kube-api-access-vcrtj") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "kube-api-access-vcrtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.347598 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a66f9850-4704-46f3-9745-bdf80caf8a98" (UID: "a66f9850-4704-46f3-9745-bdf80caf8a98"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.427759 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-run-ovn-kubernetes\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.427806 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-cni-netd\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.427862 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-ovnkube-config\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.427886 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-env-overrides\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.427902 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-etc-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428026 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-ovnkube-script-lib\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428042 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-kubelet\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428056 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-slash\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428069 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-run-netns\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428084 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-systemd-units\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428102 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428118 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-698zj\" (UniqueName: \"kubernetes.io/projected/23aa79d0-f456-425b-9557-8d109cc1aa16-kube-api-access-698zj\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428132 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-ovn\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428145 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-node-log\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428189 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-cni-bin\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428206 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/23aa79d0-f456-425b-9557-8d109cc1aa16-ovn-node-metrics-cert\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428222 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-var-lib-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428237 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428252 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-log-socket\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428271 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-systemd\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428310 4838 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428320 4838 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-slash\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428328 4838 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428336 4838 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428346 4838 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428354 4838 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-node-log\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428362 4838 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428370 4838 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66f9850-4704-46f3-9745-bdf80caf8a98-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428379 4838 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428387 4838 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428395 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcrtj\" (UniqueName: \"kubernetes.io/projected/a66f9850-4704-46f3-9745-bdf80caf8a98-kube-api-access-vcrtj\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428404 4838 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428412 4838 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428420 4838 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428427 4838 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-log-socket\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428438 4838 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428446 4838 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66f9850-4704-46f3-9745-bdf80caf8a98-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428454 4838 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428462 4838 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.428470 4838 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66f9850-4704-46f3-9745-bdf80caf8a98-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.529783 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-cni-bin\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.529863 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/23aa79d0-f456-425b-9557-8d109cc1aa16-ovn-node-metrics-cert\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.529886 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-var-lib-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.529908 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.529934 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-log-socket\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.529963 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-systemd\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.529995 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-run-ovn-kubernetes\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530024 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-cni-netd\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530049 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-ovnkube-config\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530057 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-cni-bin\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530089 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-env-overrides\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530156 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-etc-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530153 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-systemd\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530205 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-ovnkube-script-lib\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530210 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530224 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-run-netns\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530301 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-cni-netd\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530314 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-kubelet\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530243 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-run-netns\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530348 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-slash\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530380 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-systemd-units\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530417 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530448 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-698zj\" (UniqueName: \"kubernetes.io/projected/23aa79d0-f456-425b-9557-8d109cc1aa16-kube-api-access-698zj\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530476 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-ovn\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530478 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-run-ovn-kubernetes\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530501 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-node-log\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530583 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530590 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-slash\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530173 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-var-lib-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530992 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-env-overrides\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530641 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-log-socket\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530262 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-etc-openvswitch\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.530598 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-node-log\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.531197 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-run-ovn\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.531601 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-ovnkube-script-lib\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.531614 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/23aa79d0-f456-425b-9557-8d109cc1aa16-ovnkube-config\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.531691 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-host-kubelet\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.531955 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/23aa79d0-f456-425b-9557-8d109cc1aa16-systemd-units\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.536325 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/23aa79d0-f456-425b-9557-8d109cc1aa16-ovn-node-metrics-cert\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.553658 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-698zj\" (UniqueName: \"kubernetes.io/projected/23aa79d0-f456-425b-9557-8d109cc1aa16-kube-api-access-698zj\") pod \"ovnkube-node-cbtsn\" (UID: \"23aa79d0-f456-425b-9557-8d109cc1aa16\") " pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.673363 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.804258 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovnkube-controller/3.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.814019 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovn-acl-logging/0.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.815061 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-brr6g_a66f9850-4704-46f3-9745-bdf80caf8a98/ovn-controller/0.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816292 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" exitCode=0 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816344 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" exitCode=0 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816372 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" exitCode=0 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816396 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" exitCode=0 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816419 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" exitCode=0 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816438 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" exitCode=0 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816454 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" exitCode=143 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816473 4838 generic.go:334] "Generic (PLEG): container finished" podID="a66f9850-4704-46f3-9745-bdf80caf8a98" containerID="d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" exitCode=143 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816482 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816547 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816454 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816581 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816609 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816637 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816667 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816692 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816716 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816732 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816748 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816762 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816777 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816793 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816806 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816882 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816906 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816933 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816951 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816967 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816981 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.816995 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817008 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817020 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817033 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817048 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817061 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817080 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817100 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817169 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817185 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817198 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817218 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817231 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817243 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817256 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817271 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817283 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817303 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-brr6g" event={"ID":"a66f9850-4704-46f3-9745-bdf80caf8a98","Type":"ContainerDied","Data":"3e8f79c96250a7a72924c64a1f8cec3f7c67c4ee40f4070770832d7870792fe3"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817325 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817342 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817356 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817380 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817394 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817409 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817422 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817436 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817451 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817465 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.817493 4838 scope.go:117] "RemoveContainer" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.821916 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"fd1ab1ebea82b71322e971ed0b2dc82f75d982ff039e2fe3aad2b34ca049626d"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.828696 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/2.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.829578 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/1.log" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.829622 4838 generic.go:334] "Generic (PLEG): container finished" podID="ca7ddad3-e3c1-47c4-b4a8-d332f9d69227" containerID="5d011f28192f6e3f2d937b0835d48b047eb58e40d4a7df64368b680dcf332d8e" exitCode=2 Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.829653 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerDied","Data":"5d011f28192f6e3f2d937b0835d48b047eb58e40d4a7df64368b680dcf332d8e"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.829675 4838 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9"} Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.830174 4838 scope.go:117] "RemoveContainer" containerID="5d011f28192f6e3f2d937b0835d48b047eb58e40d4a7df64368b680dcf332d8e" Dec 07 09:18:16 crc kubenswrapper[4838]: E1207 09:18:16.830386 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bgkt9_openshift-multus(ca7ddad3-e3c1-47c4-b4a8-d332f9d69227)\"" pod="openshift-multus/multus-bgkt9" podUID="ca7ddad3-e3c1-47c4-b4a8-d332f9d69227" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.865614 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.904838 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-brr6g"] Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.907932 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-brr6g"] Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.929687 4838 scope.go:117] "RemoveContainer" containerID="2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.943307 4838 scope.go:117] "RemoveContainer" containerID="96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.955555 4838 scope.go:117] "RemoveContainer" containerID="d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.976043 4838 scope.go:117] "RemoveContainer" containerID="4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" Dec 07 09:18:16 crc kubenswrapper[4838]: I1207 09:18:16.993311 4838 scope.go:117] "RemoveContainer" containerID="8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.008761 4838 scope.go:117] "RemoveContainer" containerID="6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.049890 4838 scope.go:117] "RemoveContainer" containerID="d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.064792 4838 scope.go:117] "RemoveContainer" containerID="60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.077126 4838 scope.go:117] "RemoveContainer" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.077541 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": container with ID starting with ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275 not found: ID does not exist" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.077571 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} err="failed to get container status \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": rpc error: code = NotFound desc = could not find container \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": container with ID starting with ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.077613 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.077850 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": container with ID starting with 15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9 not found: ID does not exist" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.077888 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} err="failed to get container status \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": rpc error: code = NotFound desc = could not find container \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": container with ID starting with 15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.077915 4838 scope.go:117] "RemoveContainer" containerID="2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.078322 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": container with ID starting with 2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef not found: ID does not exist" containerID="2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.078357 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} err="failed to get container status \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": rpc error: code = NotFound desc = could not find container \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": container with ID starting with 2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.078373 4838 scope.go:117] "RemoveContainer" containerID="96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.078583 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": container with ID starting with 96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b not found: ID does not exist" containerID="96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.078606 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} err="failed to get container status \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": rpc error: code = NotFound desc = could not find container \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": container with ID starting with 96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.078621 4838 scope.go:117] "RemoveContainer" containerID="d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.078842 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": container with ID starting with d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837 not found: ID does not exist" containerID="d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.078866 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} err="failed to get container status \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": rpc error: code = NotFound desc = could not find container \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": container with ID starting with d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.078883 4838 scope.go:117] "RemoveContainer" containerID="4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.079141 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": container with ID starting with 4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d not found: ID does not exist" containerID="4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.079157 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} err="failed to get container status \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": rpc error: code = NotFound desc = could not find container \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": container with ID starting with 4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.079170 4838 scope.go:117] "RemoveContainer" containerID="8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.079407 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": container with ID starting with 8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7 not found: ID does not exist" containerID="8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.079431 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} err="failed to get container status \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": rpc error: code = NotFound desc = could not find container \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": container with ID starting with 8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.079447 4838 scope.go:117] "RemoveContainer" containerID="6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.080001 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": container with ID starting with 6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246 not found: ID does not exist" containerID="6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.080029 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} err="failed to get container status \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": rpc error: code = NotFound desc = could not find container \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": container with ID starting with 6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.080044 4838 scope.go:117] "RemoveContainer" containerID="d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.080432 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": container with ID starting with d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd not found: ID does not exist" containerID="d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.080526 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} err="failed to get container status \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": rpc error: code = NotFound desc = could not find container \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": container with ID starting with d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.080604 4838 scope.go:117] "RemoveContainer" containerID="60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636" Dec 07 09:18:17 crc kubenswrapper[4838]: E1207 09:18:17.080915 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": container with ID starting with 60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636 not found: ID does not exist" containerID="60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.080947 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} err="failed to get container status \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": rpc error: code = NotFound desc = could not find container \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": container with ID starting with 60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.080962 4838 scope.go:117] "RemoveContainer" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.081265 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} err="failed to get container status \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": rpc error: code = NotFound desc = could not find container \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": container with ID starting with ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.081333 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.081678 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} err="failed to get container status \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": rpc error: code = NotFound desc = could not find container \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": container with ID starting with 15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.081697 4838 scope.go:117] "RemoveContainer" containerID="2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.082261 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} err="failed to get container status \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": rpc error: code = NotFound desc = could not find container \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": container with ID starting with 2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.082331 4838 scope.go:117] "RemoveContainer" containerID="96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.082616 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} err="failed to get container status \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": rpc error: code = NotFound desc = could not find container \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": container with ID starting with 96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.082634 4838 scope.go:117] "RemoveContainer" containerID="d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.082939 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} err="failed to get container status \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": rpc error: code = NotFound desc = could not find container \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": container with ID starting with d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.083008 4838 scope.go:117] "RemoveContainer" containerID="4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.083288 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} err="failed to get container status \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": rpc error: code = NotFound desc = could not find container \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": container with ID starting with 4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.083306 4838 scope.go:117] "RemoveContainer" containerID="8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.083562 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} err="failed to get container status \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": rpc error: code = NotFound desc = could not find container \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": container with ID starting with 8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.083605 4838 scope.go:117] "RemoveContainer" containerID="6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.083883 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} err="failed to get container status \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": rpc error: code = NotFound desc = could not find container \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": container with ID starting with 6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.083904 4838 scope.go:117] "RemoveContainer" containerID="d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.084092 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} err="failed to get container status \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": rpc error: code = NotFound desc = could not find container \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": container with ID starting with d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.084115 4838 scope.go:117] "RemoveContainer" containerID="60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.084442 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} err="failed to get container status \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": rpc error: code = NotFound desc = could not find container \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": container with ID starting with 60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.084479 4838 scope.go:117] "RemoveContainer" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.084793 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} err="failed to get container status \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": rpc error: code = NotFound desc = could not find container \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": container with ID starting with ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.084836 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085160 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} err="failed to get container status \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": rpc error: code = NotFound desc = could not find container \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": container with ID starting with 15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085182 4838 scope.go:117] "RemoveContainer" containerID="2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085384 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} err="failed to get container status \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": rpc error: code = NotFound desc = could not find container \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": container with ID starting with 2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085407 4838 scope.go:117] "RemoveContainer" containerID="96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085687 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} err="failed to get container status \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": rpc error: code = NotFound desc = could not find container \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": container with ID starting with 96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085714 4838 scope.go:117] "RemoveContainer" containerID="d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085938 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} err="failed to get container status \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": rpc error: code = NotFound desc = could not find container \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": container with ID starting with d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.085959 4838 scope.go:117] "RemoveContainer" containerID="4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086140 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} err="failed to get container status \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": rpc error: code = NotFound desc = could not find container \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": container with ID starting with 4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086162 4838 scope.go:117] "RemoveContainer" containerID="8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086348 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} err="failed to get container status \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": rpc error: code = NotFound desc = could not find container \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": container with ID starting with 8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086371 4838 scope.go:117] "RemoveContainer" containerID="6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086556 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} err="failed to get container status \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": rpc error: code = NotFound desc = could not find container \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": container with ID starting with 6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086581 4838 scope.go:117] "RemoveContainer" containerID="d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086919 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} err="failed to get container status \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": rpc error: code = NotFound desc = could not find container \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": container with ID starting with d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.086939 4838 scope.go:117] "RemoveContainer" containerID="60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087155 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} err="failed to get container status \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": rpc error: code = NotFound desc = could not find container \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": container with ID starting with 60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087179 4838 scope.go:117] "RemoveContainer" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087391 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} err="failed to get container status \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": rpc error: code = NotFound desc = could not find container \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": container with ID starting with ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087413 4838 scope.go:117] "RemoveContainer" containerID="15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087651 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9"} err="failed to get container status \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": rpc error: code = NotFound desc = could not find container \"15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9\": container with ID starting with 15055429ed263ea85f2239b2040c8cc97d405b19153ecfb691b9eda3c63d12d9 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087668 4838 scope.go:117] "RemoveContainer" containerID="2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087882 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef"} err="failed to get container status \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": rpc error: code = NotFound desc = could not find container \"2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef\": container with ID starting with 2abdc1677ed0a06965a9b171b059883935bf1159107da4a3fc785ae1dab5ffef not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.087905 4838 scope.go:117] "RemoveContainer" containerID="96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088125 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b"} err="failed to get container status \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": rpc error: code = NotFound desc = could not find container \"96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b\": container with ID starting with 96155746c2cb42abd2899c6135bdd4e104010e4111f0a37815ac966fd53b3f1b not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088147 4838 scope.go:117] "RemoveContainer" containerID="d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088383 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837"} err="failed to get container status \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": rpc error: code = NotFound desc = could not find container \"d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837\": container with ID starting with d426e4c5506328d7b5632251b2b29cf2bf1159db8f04b76a477644987e2ac837 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088423 4838 scope.go:117] "RemoveContainer" containerID="4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088628 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d"} err="failed to get container status \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": rpc error: code = NotFound desc = could not find container \"4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d\": container with ID starting with 4a13d76e008861f5a29e1b139120d033ebfe82f6aed1c3f27f68658a851a791d not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088652 4838 scope.go:117] "RemoveContainer" containerID="8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088861 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7"} err="failed to get container status \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": rpc error: code = NotFound desc = could not find container \"8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7\": container with ID starting with 8486500d900b802e5c0508bd4c682d20f5f97cdffd2ab4a23a14ea67446685f7 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.088878 4838 scope.go:117] "RemoveContainer" containerID="6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.089090 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246"} err="failed to get container status \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": rpc error: code = NotFound desc = could not find container \"6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246\": container with ID starting with 6fb592492c4fae6836f6edc2619acc7484d18a00c4f4fb5a4b04874a60e97246 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.089117 4838 scope.go:117] "RemoveContainer" containerID="d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.089351 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd"} err="failed to get container status \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": rpc error: code = NotFound desc = could not find container \"d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd\": container with ID starting with d61704c4ca4e7a32313b873273e117befbb991b829d5a16268a422e0fab43efd not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.089377 4838 scope.go:117] "RemoveContainer" containerID="60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.089598 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636"} err="failed to get container status \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": rpc error: code = NotFound desc = could not find container \"60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636\": container with ID starting with 60e33b23e563035cde97ab2f12d60bdbb1cc5e527d8415ae9ae87165cd83f636 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.089638 4838 scope.go:117] "RemoveContainer" containerID="ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.089871 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275"} err="failed to get container status \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": rpc error: code = NotFound desc = could not find container \"ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275\": container with ID starting with ebcb5ce3ef0203d3ee5d4fcbe31a9471f7bfd905ab16f86dd43597d6918ec275 not found: ID does not exist" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.631264 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a66f9850-4704-46f3-9745-bdf80caf8a98" path="/var/lib/kubelet/pods/a66f9850-4704-46f3-9745-bdf80caf8a98/volumes" Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.837136 4838 generic.go:334] "Generic (PLEG): container finished" podID="23aa79d0-f456-425b-9557-8d109cc1aa16" containerID="5eb3e1dbdd212824503741d42f4846c9d88ee785a0acc7476ddee75b8733731c" exitCode=0 Dec 07 09:18:17 crc kubenswrapper[4838]: I1207 09:18:17.837170 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerDied","Data":"5eb3e1dbdd212824503741d42f4846c9d88ee785a0acc7476ddee75b8733731c"} Dec 07 09:18:18 crc kubenswrapper[4838]: I1207 09:18:18.849038 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"dcd5d898042dd86f785dfe444b920ec14a32fe96484827f0ff61c1f4d3b66b86"} Dec 07 09:18:18 crc kubenswrapper[4838]: I1207 09:18:18.849356 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"57a58f0ce4c8b923d8def5057e23a95c349641b38c86bc6f6f0635a46f7b162a"} Dec 07 09:18:18 crc kubenswrapper[4838]: I1207 09:18:18.849372 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"613e23d7d63d8628812e1d4b083081d0b7b806141b019b5ddd8ddcda3317c39e"} Dec 07 09:18:18 crc kubenswrapper[4838]: I1207 09:18:18.849384 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"bfc25d0f44a0ced78d0416105f733b717775b88f865b59c93063ad226ecf9a71"} Dec 07 09:18:18 crc kubenswrapper[4838]: I1207 09:18:18.849395 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"bd37aac63d2aca677d72c6e660e0ce791d871302b090ea1fc05874927c1f959a"} Dec 07 09:18:18 crc kubenswrapper[4838]: I1207 09:18:18.849407 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"5a65a3aed8e5cbdf40f2b0dce5aee3b8169ff44194e7b4ea21b23ccf02f692b4"} Dec 07 09:18:21 crc kubenswrapper[4838]: I1207 09:18:21.875654 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"92433d683d19212fa55589d059bb95f4d1be0992ecb766561cbe3c0d5f5ee9e3"} Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.890436 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" event={"ID":"23aa79d0-f456-425b-9557-8d109cc1aa16","Type":"ContainerStarted","Data":"e010f44211b4c6124b06d086d6a0b10fda1806f8d9869efe5925fb24a171bcc6"} Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.891001 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.891020 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.891033 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.916400 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.917339 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" podStartSLOduration=7.917323855 podStartE2EDuration="7.917323855s" podCreationTimestamp="2025-12-07 09:18:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:18:23.914890206 +0000 UTC m=+720.622209233" watchObservedRunningTime="2025-12-07 09:18:23.917323855 +0000 UTC m=+720.624642872" Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.924091 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:23 crc kubenswrapper[4838]: I1207 09:18:23.932225 4838 scope.go:117] "RemoveContainer" containerID="6ae786332ff3925be473c0763e346edf1653b7a6e9baf87a473250a875048fa9" Dec 07 09:18:24 crc kubenswrapper[4838]: I1207 09:18:24.492880 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:18:24 crc kubenswrapper[4838]: I1207 09:18:24.492946 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:18:24 crc kubenswrapper[4838]: I1207 09:18:24.899070 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/2.log" Dec 07 09:18:30 crc kubenswrapper[4838]: I1207 09:18:30.615434 4838 scope.go:117] "RemoveContainer" containerID="5d011f28192f6e3f2d937b0835d48b047eb58e40d4a7df64368b680dcf332d8e" Dec 07 09:18:30 crc kubenswrapper[4838]: E1207 09:18:30.616613 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bgkt9_openshift-multus(ca7ddad3-e3c1-47c4-b4a8-d332f9d69227)\"" pod="openshift-multus/multus-bgkt9" podUID="ca7ddad3-e3c1-47c4-b4a8-d332f9d69227" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.041890 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl"] Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.043595 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.045908 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.061560 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl"] Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.201546 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.201855 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rps76\" (UniqueName: \"kubernetes.io/projected/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-kube-api-access-rps76\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.201959 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.303007 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.303080 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rps76\" (UniqueName: \"kubernetes.io/projected/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-kube-api-access-rps76\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.303111 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.303453 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.303473 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.328716 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rps76\" (UniqueName: \"kubernetes.io/projected/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-kube-api-access-rps76\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: I1207 09:18:44.370310 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: E1207 09:18:44.393380 4838 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(cd713e03863184c236dccb8070ba45faf82bcc7d61aecfc1928083b221f036c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 09:18:44 crc kubenswrapper[4838]: E1207 09:18:44.393503 4838 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(cd713e03863184c236dccb8070ba45faf82bcc7d61aecfc1928083b221f036c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: E1207 09:18:44.393574 4838 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(cd713e03863184c236dccb8070ba45faf82bcc7d61aecfc1928083b221f036c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:44 crc kubenswrapper[4838]: E1207 09:18:44.393673 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace(1487e6d4-4c48-4b9a-a814-fa6c0431ae7c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace(1487e6d4-4c48-4b9a-a814-fa6c0431ae7c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(cd713e03863184c236dccb8070ba45faf82bcc7d61aecfc1928083b221f036c6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" Dec 07 09:18:45 crc kubenswrapper[4838]: I1207 09:18:45.038110 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:45 crc kubenswrapper[4838]: I1207 09:18:45.039118 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:45 crc kubenswrapper[4838]: E1207 09:18:45.089427 4838 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(38cbcf61096681633f0ee33936b2610de313065024fc1fc223140f9b290515d3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 09:18:45 crc kubenswrapper[4838]: E1207 09:18:45.089526 4838 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(38cbcf61096681633f0ee33936b2610de313065024fc1fc223140f9b290515d3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:45 crc kubenswrapper[4838]: E1207 09:18:45.089581 4838 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(38cbcf61096681633f0ee33936b2610de313065024fc1fc223140f9b290515d3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:45 crc kubenswrapper[4838]: E1207 09:18:45.089768 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace(1487e6d4-4c48-4b9a-a814-fa6c0431ae7c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace(1487e6d4-4c48-4b9a-a814-fa6c0431ae7c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_openshift-marketplace_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c_0(38cbcf61096681633f0ee33936b2610de313065024fc1fc223140f9b290515d3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" Dec 07 09:18:45 crc kubenswrapper[4838]: I1207 09:18:45.614310 4838 scope.go:117] "RemoveContainer" containerID="5d011f28192f6e3f2d937b0835d48b047eb58e40d4a7df64368b680dcf332d8e" Dec 07 09:18:46 crc kubenswrapper[4838]: I1207 09:18:46.044670 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bgkt9_ca7ddad3-e3c1-47c4-b4a8-d332f9d69227/kube-multus/2.log" Dec 07 09:18:46 crc kubenswrapper[4838]: I1207 09:18:46.044720 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bgkt9" event={"ID":"ca7ddad3-e3c1-47c4-b4a8-d332f9d69227","Type":"ContainerStarted","Data":"8a49aeca9acde7d38f1674e0b0dd65ea7ca5e9a711c9a3645674d4211470e6a6"} Dec 07 09:18:46 crc kubenswrapper[4838]: I1207 09:18:46.705412 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cbtsn" Dec 07 09:18:54 crc kubenswrapper[4838]: I1207 09:18:54.493665 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:18:54 crc kubenswrapper[4838]: I1207 09:18:54.494398 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:18:59 crc kubenswrapper[4838]: I1207 09:18:59.614518 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:59 crc kubenswrapper[4838]: I1207 09:18:59.615531 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:18:59 crc kubenswrapper[4838]: I1207 09:18:59.856864 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl"] Dec 07 09:19:00 crc kubenswrapper[4838]: I1207 09:19:00.137686 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" event={"ID":"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c","Type":"ContainerStarted","Data":"2b482caf38177a1b440152389ffa0272f3c61ba67c77a56fc7b6acf2eb49ed13"} Dec 07 09:19:00 crc kubenswrapper[4838]: I1207 09:19:00.138300 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" event={"ID":"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c","Type":"ContainerStarted","Data":"99b29b70dc23a867900705a1c7a68c8ff324ee676999861b0ca244fd668519d7"} Dec 07 09:19:01 crc kubenswrapper[4838]: I1207 09:19:01.147260 4838 generic.go:334] "Generic (PLEG): container finished" podID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerID="2b482caf38177a1b440152389ffa0272f3c61ba67c77a56fc7b6acf2eb49ed13" exitCode=0 Dec 07 09:19:01 crc kubenswrapper[4838]: I1207 09:19:01.147323 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" event={"ID":"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c","Type":"ContainerDied","Data":"2b482caf38177a1b440152389ffa0272f3c61ba67c77a56fc7b6acf2eb49ed13"} Dec 07 09:19:02 crc kubenswrapper[4838]: I1207 09:19:02.542449 4838 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.160752 4838 generic.go:334] "Generic (PLEG): container finished" podID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerID="2b7be8fac357b0e963aaec30f9aacbb9806704b383f8f2df0a5784b0bd836eea" exitCode=0 Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.160798 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" event={"ID":"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c","Type":"ContainerDied","Data":"2b7be8fac357b0e963aaec30f9aacbb9806704b383f8f2df0a5784b0bd836eea"} Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.425062 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fz2zx"] Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.426636 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.444375 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fz2zx"] Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.555951 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-catalog-content\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.556471 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-utilities\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.556645 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6kgz\" (UniqueName: \"kubernetes.io/projected/5ee59b7f-9315-451e-91ad-034f28612374-kube-api-access-w6kgz\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.657633 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6kgz\" (UniqueName: \"kubernetes.io/projected/5ee59b7f-9315-451e-91ad-034f28612374-kube-api-access-w6kgz\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.658148 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-catalog-content\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.658287 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-utilities\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.658721 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-utilities\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.658726 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-catalog-content\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.676426 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6kgz\" (UniqueName: \"kubernetes.io/projected/5ee59b7f-9315-451e-91ad-034f28612374-kube-api-access-w6kgz\") pod \"redhat-operators-fz2zx\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:03 crc kubenswrapper[4838]: I1207 09:19:03.755396 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:04 crc kubenswrapper[4838]: I1207 09:19:04.003873 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fz2zx"] Dec 07 09:19:04 crc kubenswrapper[4838]: I1207 09:19:04.167396 4838 generic.go:334] "Generic (PLEG): container finished" podID="5ee59b7f-9315-451e-91ad-034f28612374" containerID="f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e" exitCode=0 Dec 07 09:19:04 crc kubenswrapper[4838]: I1207 09:19:04.167434 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fz2zx" event={"ID":"5ee59b7f-9315-451e-91ad-034f28612374","Type":"ContainerDied","Data":"f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e"} Dec 07 09:19:04 crc kubenswrapper[4838]: I1207 09:19:04.167470 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fz2zx" event={"ID":"5ee59b7f-9315-451e-91ad-034f28612374","Type":"ContainerStarted","Data":"3f41d80ef125d66869897100eeba2652eb446b703cf80ae6a26959bb474eefb6"} Dec 07 09:19:04 crc kubenswrapper[4838]: I1207 09:19:04.169741 4838 generic.go:334] "Generic (PLEG): container finished" podID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerID="a3a6bd34026984d7d82a0a94bdd7d663f7bb8ee24a60d7f7d628a5773a1290ba" exitCode=0 Dec 07 09:19:04 crc kubenswrapper[4838]: I1207 09:19:04.169786 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" event={"ID":"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c","Type":"ContainerDied","Data":"a3a6bd34026984d7d82a0a94bdd7d663f7bb8ee24a60d7f7d628a5773a1290ba"} Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.180867 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fz2zx" event={"ID":"5ee59b7f-9315-451e-91ad-034f28612374","Type":"ContainerStarted","Data":"c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1"} Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.440924 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.580707 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-util\") pod \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.581001 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-bundle\") pod \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.581610 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rps76\" (UniqueName: \"kubernetes.io/projected/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-kube-api-access-rps76\") pod \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\" (UID: \"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c\") " Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.583950 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-bundle" (OuterVolumeSpecName: "bundle") pod "1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" (UID: "1487e6d4-4c48-4b9a-a814-fa6c0431ae7c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.595124 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-kube-api-access-rps76" (OuterVolumeSpecName: "kube-api-access-rps76") pod "1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" (UID: "1487e6d4-4c48-4b9a-a814-fa6c0431ae7c"). InnerVolumeSpecName "kube-api-access-rps76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.606394 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-util" (OuterVolumeSpecName: "util") pod "1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" (UID: "1487e6d4-4c48-4b9a-a814-fa6c0431ae7c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.683706 4838 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-util\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.683738 4838 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:05 crc kubenswrapper[4838]: I1207 09:19:05.683750 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rps76\" (UniqueName: \"kubernetes.io/projected/1487e6d4-4c48-4b9a-a814-fa6c0431ae7c-kube-api-access-rps76\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:06 crc kubenswrapper[4838]: I1207 09:19:06.186971 4838 generic.go:334] "Generic (PLEG): container finished" podID="5ee59b7f-9315-451e-91ad-034f28612374" containerID="c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1" exitCode=0 Dec 07 09:19:06 crc kubenswrapper[4838]: I1207 09:19:06.187083 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fz2zx" event={"ID":"5ee59b7f-9315-451e-91ad-034f28612374","Type":"ContainerDied","Data":"c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1"} Dec 07 09:19:06 crc kubenswrapper[4838]: I1207 09:19:06.190249 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" event={"ID":"1487e6d4-4c48-4b9a-a814-fa6c0431ae7c","Type":"ContainerDied","Data":"99b29b70dc23a867900705a1c7a68c8ff324ee676999861b0ca244fd668519d7"} Dec 07 09:19:06 crc kubenswrapper[4838]: I1207 09:19:06.190359 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99b29b70dc23a867900705a1c7a68c8ff324ee676999861b0ca244fd668519d7" Dec 07 09:19:06 crc kubenswrapper[4838]: I1207 09:19:06.190335 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl" Dec 07 09:19:07 crc kubenswrapper[4838]: I1207 09:19:07.202553 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fz2zx" event={"ID":"5ee59b7f-9315-451e-91ad-034f28612374","Type":"ContainerStarted","Data":"a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3"} Dec 07 09:19:07 crc kubenswrapper[4838]: I1207 09:19:07.228766 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fz2zx" podStartSLOduration=1.795853663 podStartE2EDuration="4.22874782s" podCreationTimestamp="2025-12-07 09:19:03 +0000 UTC" firstStartedPulling="2025-12-07 09:19:04.168669151 +0000 UTC m=+760.875988158" lastFinishedPulling="2025-12-07 09:19:06.601563258 +0000 UTC m=+763.308882315" observedRunningTime="2025-12-07 09:19:07.225329974 +0000 UTC m=+763.932648991" watchObservedRunningTime="2025-12-07 09:19:07.22874782 +0000 UTC m=+763.936066847" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.987739 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm"] Dec 07 09:19:09 crc kubenswrapper[4838]: E1207 09:19:09.989040 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerName="pull" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.989147 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerName="pull" Dec 07 09:19:09 crc kubenswrapper[4838]: E1207 09:19:09.989223 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerName="util" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.989295 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerName="util" Dec 07 09:19:09 crc kubenswrapper[4838]: E1207 09:19:09.989376 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerName="extract" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.989450 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerName="extract" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.989640 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="1487e6d4-4c48-4b9a-a814-fa6c0431ae7c" containerName="extract" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.990174 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.995129 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.995502 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 07 09:19:09 crc kubenswrapper[4838]: I1207 09:19:09.995627 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bgk8k" Dec 07 09:19:10 crc kubenswrapper[4838]: I1207 09:19:10.005988 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm"] Dec 07 09:19:10 crc kubenswrapper[4838]: I1207 09:19:10.140960 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb95m\" (UniqueName: \"kubernetes.io/projected/e1df0bd8-ff98-4094-92b6-23f266883f8a-kube-api-access-hb95m\") pod \"nmstate-operator-5b5b58f5c8-4hcdm\" (UID: \"e1df0bd8-ff98-4094-92b6-23f266883f8a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" Dec 07 09:19:10 crc kubenswrapper[4838]: I1207 09:19:10.241919 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb95m\" (UniqueName: \"kubernetes.io/projected/e1df0bd8-ff98-4094-92b6-23f266883f8a-kube-api-access-hb95m\") pod \"nmstate-operator-5b5b58f5c8-4hcdm\" (UID: \"e1df0bd8-ff98-4094-92b6-23f266883f8a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" Dec 07 09:19:10 crc kubenswrapper[4838]: I1207 09:19:10.259531 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb95m\" (UniqueName: \"kubernetes.io/projected/e1df0bd8-ff98-4094-92b6-23f266883f8a-kube-api-access-hb95m\") pod \"nmstate-operator-5b5b58f5c8-4hcdm\" (UID: \"e1df0bd8-ff98-4094-92b6-23f266883f8a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" Dec 07 09:19:10 crc kubenswrapper[4838]: I1207 09:19:10.347295 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" Dec 07 09:19:10 crc kubenswrapper[4838]: I1207 09:19:10.560694 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm"] Dec 07 09:19:11 crc kubenswrapper[4838]: I1207 09:19:11.226528 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" event={"ID":"e1df0bd8-ff98-4094-92b6-23f266883f8a","Type":"ContainerStarted","Data":"58db5deaf6ae2951cd837b3c129deee0bda499a67203f63fbba3b2f9c5fe37db"} Dec 07 09:19:13 crc kubenswrapper[4838]: I1207 09:19:13.756201 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:13 crc kubenswrapper[4838]: I1207 09:19:13.756519 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:13 crc kubenswrapper[4838]: I1207 09:19:13.791291 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:14 crc kubenswrapper[4838]: I1207 09:19:14.244914 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" event={"ID":"e1df0bd8-ff98-4094-92b6-23f266883f8a","Type":"ContainerStarted","Data":"d5ca92606f30a8e6d60195b5374d963b46794e71e2e3c28d50e7554f3c5b680a"} Dec 07 09:19:14 crc kubenswrapper[4838]: I1207 09:19:14.262751 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-4hcdm" podStartSLOduration=1.7928972779999999 podStartE2EDuration="5.262731004s" podCreationTimestamp="2025-12-07 09:19:09 +0000 UTC" firstStartedPulling="2025-12-07 09:19:10.574892488 +0000 UTC m=+767.282211505" lastFinishedPulling="2025-12-07 09:19:14.044726214 +0000 UTC m=+770.752045231" observedRunningTime="2025-12-07 09:19:14.262066565 +0000 UTC m=+770.969385602" watchObservedRunningTime="2025-12-07 09:19:14.262731004 +0000 UTC m=+770.970050021" Dec 07 09:19:14 crc kubenswrapper[4838]: I1207 09:19:14.300140 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.252639 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.253463 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.256090 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-l78b7" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.270311 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.271051 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.272756 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.281610 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-twb76"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.282262 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.303232 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.311545 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.408158 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.409214 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.413722 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.413902 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.414273 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-s4dlt" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.415714 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-ovs-socket\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.415864 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-dbus-socket\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.415966 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-nmstate-lock\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.416043 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvhjb\" (UniqueName: \"kubernetes.io/projected/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-kube-api-access-zvhjb\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.416114 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw4bz\" (UniqueName: \"kubernetes.io/projected/9d7f7de9-0d6e-4bf5-83c0-5f703cb433af-kube-api-access-dw4bz\") pod \"nmstate-webhook-5f6d4c5ccb-jcgsg\" (UID: \"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.416185 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9d7f7de9-0d6e-4bf5-83c0-5f703cb433af-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jcgsg\" (UID: \"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.416260 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs9gs\" (UniqueName: \"kubernetes.io/projected/11834459-a487-4c2f-9521-add0e1726e9d-kube-api-access-vs9gs\") pod \"nmstate-metrics-7f946cbc9-p58hh\" (UID: \"11834459-a487-4c2f-9521-add0e1726e9d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.417568 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517531 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/013a32de-8cd7-44da-b292-5948ab0bb6c7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517591 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-ovs-socket\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517623 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvhvv\" (UniqueName: \"kubernetes.io/projected/013a32de-8cd7-44da-b292-5948ab0bb6c7-kube-api-access-vvhvv\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517651 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-dbus-socket\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517697 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-nmstate-lock\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517730 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvhjb\" (UniqueName: \"kubernetes.io/projected/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-kube-api-access-zvhjb\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517756 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw4bz\" (UniqueName: \"kubernetes.io/projected/9d7f7de9-0d6e-4bf5-83c0-5f703cb433af-kube-api-access-dw4bz\") pod \"nmstate-webhook-5f6d4c5ccb-jcgsg\" (UID: \"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517777 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/013a32de-8cd7-44da-b292-5948ab0bb6c7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517803 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9d7f7de9-0d6e-4bf5-83c0-5f703cb433af-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jcgsg\" (UID: \"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.517844 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs9gs\" (UniqueName: \"kubernetes.io/projected/11834459-a487-4c2f-9521-add0e1726e9d-kube-api-access-vs9gs\") pod \"nmstate-metrics-7f946cbc9-p58hh\" (UID: \"11834459-a487-4c2f-9521-add0e1726e9d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.518385 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-ovs-socket\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.518725 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-dbus-socket\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.518864 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-nmstate-lock\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.537761 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9d7f7de9-0d6e-4bf5-83c0-5f703cb433af-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jcgsg\" (UID: \"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.541249 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs9gs\" (UniqueName: \"kubernetes.io/projected/11834459-a487-4c2f-9521-add0e1726e9d-kube-api-access-vs9gs\") pod \"nmstate-metrics-7f946cbc9-p58hh\" (UID: \"11834459-a487-4c2f-9521-add0e1726e9d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.541512 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvhjb\" (UniqueName: \"kubernetes.io/projected/a130a77c-66e9-48b2-a9c9-36c44b7b8eec-kube-api-access-zvhjb\") pod \"nmstate-handler-twb76\" (UID: \"a130a77c-66e9-48b2-a9c9-36c44b7b8eec\") " pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.544480 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw4bz\" (UniqueName: \"kubernetes.io/projected/9d7f7de9-0d6e-4bf5-83c0-5f703cb433af-kube-api-access-dw4bz\") pod \"nmstate-webhook-5f6d4c5ccb-jcgsg\" (UID: \"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.578584 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.604906 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.621524 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/013a32de-8cd7-44da-b292-5948ab0bb6c7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.621586 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/013a32de-8cd7-44da-b292-5948ab0bb6c7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.621607 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvhvv\" (UniqueName: \"kubernetes.io/projected/013a32de-8cd7-44da-b292-5948ab0bb6c7-kube-api-access-vvhvv\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.622565 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/013a32de-8cd7-44da-b292-5948ab0bb6c7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.627086 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.627834 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/013a32de-8cd7-44da-b292-5948ab0bb6c7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.650913 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7cfc4f8755-9b2fd"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.651546 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.653203 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvhvv\" (UniqueName: \"kubernetes.io/projected/013a32de-8cd7-44da-b292-5948ab0bb6c7-kube-api-access-vvhvv\") pod \"nmstate-console-plugin-7fbb5f6569-lzwd2\" (UID: \"013a32de-8cd7-44da-b292-5948ab0bb6c7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.676503 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cfc4f8755-9b2fd"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.723055 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-trusted-ca-bundle\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.723102 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-service-ca\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.723141 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-serving-cert\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.723170 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-oauth-config\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.723202 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-config\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.723219 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-oauth-serving-cert\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.723232 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78fg6\" (UniqueName: \"kubernetes.io/projected/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-kube-api-access-78fg6\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.733307 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.824967 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-serving-cert\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.825316 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-oauth-config\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.825352 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-config\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.825372 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-oauth-serving-cert\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.825387 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78fg6\" (UniqueName: \"kubernetes.io/projected/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-kube-api-access-78fg6\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.825408 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-trusted-ca-bundle\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.825430 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-service-ca\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.826277 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-service-ca\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.827235 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-oauth-serving-cert\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.827294 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-config\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.828005 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-trusted-ca-bundle\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.829564 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-oauth-config\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.829622 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-console-serving-cert\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.847758 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78fg6\" (UniqueName: \"kubernetes.io/projected/25a7d8e4-31e0-4717-b29c-44ec2410e3bf-kube-api-access-78fg6\") pod \"console-7cfc4f8755-9b2fd\" (UID: \"25a7d8e4-31e0-4717-b29c-44ec2410e3bf\") " pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.863362 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.899948 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh"] Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.986127 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:15 crc kubenswrapper[4838]: I1207 09:19:15.995021 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2"] Dec 07 09:19:16 crc kubenswrapper[4838]: W1207 09:19:16.001428 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod013a32de_8cd7_44da_b292_5948ab0bb6c7.slice/crio-c54b35ef05243336fc6cb09568bb0f31d3739ee88b3e8231eb8ddf3b05ec8385 WatchSource:0}: Error finding container c54b35ef05243336fc6cb09568bb0f31d3739ee88b3e8231eb8ddf3b05ec8385: Status 404 returned error can't find the container with id c54b35ef05243336fc6cb09568bb0f31d3739ee88b3e8231eb8ddf3b05ec8385 Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.178772 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cfc4f8755-9b2fd"] Dec 07 09:19:16 crc kubenswrapper[4838]: W1207 09:19:16.184371 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25a7d8e4_31e0_4717_b29c_44ec2410e3bf.slice/crio-93bdbbc125858dce4ec414dcc75d52bdb2ef328fe2334c0c9358190cddff1544 WatchSource:0}: Error finding container 93bdbbc125858dce4ec414dcc75d52bdb2ef328fe2334c0c9358190cddff1544: Status 404 returned error can't find the container with id 93bdbbc125858dce4ec414dcc75d52bdb2ef328fe2334c0c9358190cddff1544 Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.197385 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fz2zx"] Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.254821 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-twb76" event={"ID":"a130a77c-66e9-48b2-a9c9-36c44b7b8eec","Type":"ContainerStarted","Data":"34c3b5f8ebc632c04a2a414131f3e51b5ea0e110053e32a0a1cbce017996ea50"} Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.255885 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" event={"ID":"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af","Type":"ContainerStarted","Data":"e52b7a2bae658a95d70a08d04157b54b51405b7afbde8377876e74b31859ba61"} Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.257053 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cfc4f8755-9b2fd" event={"ID":"25a7d8e4-31e0-4717-b29c-44ec2410e3bf","Type":"ContainerStarted","Data":"93bdbbc125858dce4ec414dcc75d52bdb2ef328fe2334c0c9358190cddff1544"} Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.259250 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" event={"ID":"013a32de-8cd7-44da-b292-5948ab0bb6c7","Type":"ContainerStarted","Data":"c54b35ef05243336fc6cb09568bb0f31d3739ee88b3e8231eb8ddf3b05ec8385"} Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.262078 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" event={"ID":"11834459-a487-4c2f-9521-add0e1726e9d","Type":"ContainerStarted","Data":"871ff4d034799d3281b0a8a2128a6e6d33185611bf998fd635e743d379128f6c"} Dec 07 09:19:16 crc kubenswrapper[4838]: I1207 09:19:16.262174 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fz2zx" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="registry-server" containerID="cri-o://a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3" gracePeriod=2 Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.122947 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.243864 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-utilities\") pod \"5ee59b7f-9315-451e-91ad-034f28612374\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.244023 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-catalog-content\") pod \"5ee59b7f-9315-451e-91ad-034f28612374\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.244059 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6kgz\" (UniqueName: \"kubernetes.io/projected/5ee59b7f-9315-451e-91ad-034f28612374-kube-api-access-w6kgz\") pod \"5ee59b7f-9315-451e-91ad-034f28612374\" (UID: \"5ee59b7f-9315-451e-91ad-034f28612374\") " Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.244795 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-utilities" (OuterVolumeSpecName: "utilities") pod "5ee59b7f-9315-451e-91ad-034f28612374" (UID: "5ee59b7f-9315-451e-91ad-034f28612374"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.249209 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee59b7f-9315-451e-91ad-034f28612374-kube-api-access-w6kgz" (OuterVolumeSpecName: "kube-api-access-w6kgz") pod "5ee59b7f-9315-451e-91ad-034f28612374" (UID: "5ee59b7f-9315-451e-91ad-034f28612374"). InnerVolumeSpecName "kube-api-access-w6kgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.281386 4838 generic.go:334] "Generic (PLEG): container finished" podID="5ee59b7f-9315-451e-91ad-034f28612374" containerID="a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3" exitCode=0 Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.281459 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fz2zx" event={"ID":"5ee59b7f-9315-451e-91ad-034f28612374","Type":"ContainerDied","Data":"a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3"} Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.281487 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fz2zx" event={"ID":"5ee59b7f-9315-451e-91ad-034f28612374","Type":"ContainerDied","Data":"3f41d80ef125d66869897100eeba2652eb446b703cf80ae6a26959bb474eefb6"} Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.281466 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fz2zx" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.281503 4838 scope.go:117] "RemoveContainer" containerID="a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.284154 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cfc4f8755-9b2fd" event={"ID":"25a7d8e4-31e0-4717-b29c-44ec2410e3bf","Type":"ContainerStarted","Data":"6952a0b5a69a88f3d3b1c6710eac8ca67e72b549c0fcd796626957f7c9093076"} Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.307410 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7cfc4f8755-9b2fd" podStartSLOduration=2.307391727 podStartE2EDuration="2.307391727s" podCreationTimestamp="2025-12-07 09:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:19:17.299107384 +0000 UTC m=+774.006426401" watchObservedRunningTime="2025-12-07 09:19:17.307391727 +0000 UTC m=+774.014710744" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.317518 4838 scope.go:117] "RemoveContainer" containerID="c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.346949 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6kgz\" (UniqueName: \"kubernetes.io/projected/5ee59b7f-9315-451e-91ad-034f28612374-kube-api-access-w6kgz\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.346975 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.350508 4838 scope.go:117] "RemoveContainer" containerID="f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.357980 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ee59b7f-9315-451e-91ad-034f28612374" (UID: "5ee59b7f-9315-451e-91ad-034f28612374"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.367357 4838 scope.go:117] "RemoveContainer" containerID="a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3" Dec 07 09:19:17 crc kubenswrapper[4838]: E1207 09:19:17.367746 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3\": container with ID starting with a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3 not found: ID does not exist" containerID="a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.367785 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3"} err="failed to get container status \"a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3\": rpc error: code = NotFound desc = could not find container \"a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3\": container with ID starting with a1fc933d42a35d35ac2171f2f473375ac6bc126e4ca68f158478a8b4c0f776b3 not found: ID does not exist" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.367804 4838 scope.go:117] "RemoveContainer" containerID="c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1" Dec 07 09:19:17 crc kubenswrapper[4838]: E1207 09:19:17.368532 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1\": container with ID starting with c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1 not found: ID does not exist" containerID="c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.368554 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1"} err="failed to get container status \"c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1\": rpc error: code = NotFound desc = could not find container \"c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1\": container with ID starting with c76f9a339435b5bb9e4cb042f584d60dcb199331ecb749ff550a3f5aa2a756a1 not found: ID does not exist" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.368570 4838 scope.go:117] "RemoveContainer" containerID="f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e" Dec 07 09:19:17 crc kubenswrapper[4838]: E1207 09:19:17.368853 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e\": container with ID starting with f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e not found: ID does not exist" containerID="f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.368873 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e"} err="failed to get container status \"f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e\": rpc error: code = NotFound desc = could not find container \"f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e\": container with ID starting with f6780cc9eab6a546f2d776f9ebccbe46bdbb4ff98513cb7bdc44940c9cf02e3e not found: ID does not exist" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.447939 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee59b7f-9315-451e-91ad-034f28612374-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.610354 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fz2zx"] Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.613578 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fz2zx"] Dec 07 09:19:17 crc kubenswrapper[4838]: I1207 09:19:17.622300 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ee59b7f-9315-451e-91ad-034f28612374" path="/var/lib/kubelet/pods/5ee59b7f-9315-451e-91ad-034f28612374/volumes" Dec 07 09:19:19 crc kubenswrapper[4838]: I1207 09:19:19.302397 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" event={"ID":"11834459-a487-4c2f-9521-add0e1726e9d","Type":"ContainerStarted","Data":"f69b7b4b87f5e74cd2587b1945b2234fab64e99e5ce0ea47038a1bc7a8cc31d5"} Dec 07 09:19:19 crc kubenswrapper[4838]: I1207 09:19:19.304313 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" event={"ID":"013a32de-8cd7-44da-b292-5948ab0bb6c7","Type":"ContainerStarted","Data":"ef50125e372a3a35eac54a9faa7808f33b6621f992e75ac3c72ecde4560abf7d"} Dec 07 09:19:19 crc kubenswrapper[4838]: I1207 09:19:19.305870 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-twb76" event={"ID":"a130a77c-66e9-48b2-a9c9-36c44b7b8eec","Type":"ContainerStarted","Data":"e1a64430866efa6def34da435fe4712ee0331afff1df23525c9c3a9f257b3014"} Dec 07 09:19:19 crc kubenswrapper[4838]: I1207 09:19:19.306031 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:19 crc kubenswrapper[4838]: I1207 09:19:19.326798 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lzwd2" podStartSLOduration=1.661187279 podStartE2EDuration="4.326778089s" podCreationTimestamp="2025-12-07 09:19:15 +0000 UTC" firstStartedPulling="2025-12-07 09:19:16.005167524 +0000 UTC m=+772.712486541" lastFinishedPulling="2025-12-07 09:19:18.670758324 +0000 UTC m=+775.378077351" observedRunningTime="2025-12-07 09:19:19.325437582 +0000 UTC m=+776.032756629" watchObservedRunningTime="2025-12-07 09:19:19.326778089 +0000 UTC m=+776.034097116" Dec 07 09:19:19 crc kubenswrapper[4838]: I1207 09:19:19.362448 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-twb76" podStartSLOduration=1.355238869 podStartE2EDuration="4.362419295s" podCreationTimestamp="2025-12-07 09:19:15 +0000 UTC" firstStartedPulling="2025-12-07 09:19:15.66436357 +0000 UTC m=+772.371682587" lastFinishedPulling="2025-12-07 09:19:18.671543986 +0000 UTC m=+775.378863013" observedRunningTime="2025-12-07 09:19:19.350356845 +0000 UTC m=+776.057675912" watchObservedRunningTime="2025-12-07 09:19:19.362419295 +0000 UTC m=+776.069738352" Dec 07 09:19:20 crc kubenswrapper[4838]: I1207 09:19:20.314045 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" event={"ID":"9d7f7de9-0d6e-4bf5-83c0-5f703cb433af","Type":"ContainerStarted","Data":"300aef4d8bba536a6a696352906cd230713339d8976ed41f792399d7b3c228cf"} Dec 07 09:19:20 crc kubenswrapper[4838]: I1207 09:19:20.314390 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:20 crc kubenswrapper[4838]: I1207 09:19:20.341617 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" podStartSLOduration=1.764914434 podStartE2EDuration="5.341587515s" podCreationTimestamp="2025-12-07 09:19:15 +0000 UTC" firstStartedPulling="2025-12-07 09:19:15.875359722 +0000 UTC m=+772.582678739" lastFinishedPulling="2025-12-07 09:19:19.452032793 +0000 UTC m=+776.159351820" observedRunningTime="2025-12-07 09:19:20.332596952 +0000 UTC m=+777.039915969" watchObservedRunningTime="2025-12-07 09:19:20.341587515 +0000 UTC m=+777.048906532" Dec 07 09:19:21 crc kubenswrapper[4838]: I1207 09:19:21.321523 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" event={"ID":"11834459-a487-4c2f-9521-add0e1726e9d","Type":"ContainerStarted","Data":"3dd1041640889a59922225f101010561cbc03d583303c0eef1c19f816c42b934"} Dec 07 09:19:21 crc kubenswrapper[4838]: I1207 09:19:21.350456 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p58hh" podStartSLOduration=1.293489546 podStartE2EDuration="6.350427573s" podCreationTimestamp="2025-12-07 09:19:15 +0000 UTC" firstStartedPulling="2025-12-07 09:19:15.922836451 +0000 UTC m=+772.630155478" lastFinishedPulling="2025-12-07 09:19:20.979774488 +0000 UTC m=+777.687093505" observedRunningTime="2025-12-07 09:19:21.350091334 +0000 UTC m=+778.057410441" watchObservedRunningTime="2025-12-07 09:19:21.350427573 +0000 UTC m=+778.057746620" Dec 07 09:19:24 crc kubenswrapper[4838]: I1207 09:19:24.493768 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:19:24 crc kubenswrapper[4838]: I1207 09:19:24.494420 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:19:24 crc kubenswrapper[4838]: I1207 09:19:24.494501 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:19:24 crc kubenswrapper[4838]: I1207 09:19:24.495536 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e006964b42caddec4f66d1cf7932894d0547718b0bf611546057b7159422840"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:19:24 crc kubenswrapper[4838]: I1207 09:19:24.495668 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://4e006964b42caddec4f66d1cf7932894d0547718b0bf611546057b7159422840" gracePeriod=600 Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.355987 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="4e006964b42caddec4f66d1cf7932894d0547718b0bf611546057b7159422840" exitCode=0 Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.356042 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"4e006964b42caddec4f66d1cf7932894d0547718b0bf611546057b7159422840"} Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.356427 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"b20dfe6b2c79e165940e9237c72516b3c8e4bccb48b171f998bf1346e1ad6571"} Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.356459 4838 scope.go:117] "RemoveContainer" containerID="f5608e3e5fc8b2067546b7fff7ad2e9ceab4038aa8bfb537b0261a0f743498fd" Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.661786 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-twb76" Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.987173 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.987242 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:25 crc kubenswrapper[4838]: I1207 09:19:25.995473 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:26 crc kubenswrapper[4838]: I1207 09:19:26.377414 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7cfc4f8755-9b2fd" Dec 07 09:19:26 crc kubenswrapper[4838]: I1207 09:19:26.474440 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xrdpm"] Dec 07 09:19:35 crc kubenswrapper[4838]: I1207 09:19:35.641436 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jcgsg" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.282859 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd"] Dec 07 09:19:47 crc kubenswrapper[4838]: E1207 09:19:47.283875 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="extract-content" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.283891 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="extract-content" Dec 07 09:19:47 crc kubenswrapper[4838]: E1207 09:19:47.283900 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="registry-server" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.283908 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="registry-server" Dec 07 09:19:47 crc kubenswrapper[4838]: E1207 09:19:47.283930 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="extract-utilities" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.283939 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="extract-utilities" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.284058 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee59b7f-9315-451e-91ad-034f28612374" containerName="registry-server" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.285055 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.287154 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.299755 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd"] Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.394386 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.394445 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skg2m\" (UniqueName: \"kubernetes.io/projected/36061897-0432-4e69-add8-2a3560a583b5-kube-api-access-skg2m\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.394488 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.495667 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.495767 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skg2m\" (UniqueName: \"kubernetes.io/projected/36061897-0432-4e69-add8-2a3560a583b5-kube-api-access-skg2m\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.495895 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.496254 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.496293 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.523680 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skg2m\" (UniqueName: \"kubernetes.io/projected/36061897-0432-4e69-add8-2a3560a583b5-kube-api-access-skg2m\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.599146 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:47 crc kubenswrapper[4838]: I1207 09:19:47.793431 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd"] Dec 07 09:19:48 crc kubenswrapper[4838]: I1207 09:19:48.510145 4838 generic.go:334] "Generic (PLEG): container finished" podID="36061897-0432-4e69-add8-2a3560a583b5" containerID="f6a2604503a3ee4df1de8ae7f30a8da8a86539f2ff1a3418ca44cc462bf60202" exitCode=0 Dec 07 09:19:48 crc kubenswrapper[4838]: I1207 09:19:48.510432 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" event={"ID":"36061897-0432-4e69-add8-2a3560a583b5","Type":"ContainerDied","Data":"f6a2604503a3ee4df1de8ae7f30a8da8a86539f2ff1a3418ca44cc462bf60202"} Dec 07 09:19:48 crc kubenswrapper[4838]: I1207 09:19:48.511665 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" event={"ID":"36061897-0432-4e69-add8-2a3560a583b5","Type":"ContainerStarted","Data":"1dc779c74d22483729c83a91505f981f3843a131f60ae0cbdfef005e0ca07ce1"} Dec 07 09:19:50 crc kubenswrapper[4838]: I1207 09:19:50.532344 4838 generic.go:334] "Generic (PLEG): container finished" podID="36061897-0432-4e69-add8-2a3560a583b5" containerID="d146d9cfd23e760f6d22a25a7b7b9318996a977eb18d367a0f04674eca0defc9" exitCode=0 Dec 07 09:19:50 crc kubenswrapper[4838]: I1207 09:19:50.532557 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" event={"ID":"36061897-0432-4e69-add8-2a3560a583b5","Type":"ContainerDied","Data":"d146d9cfd23e760f6d22a25a7b7b9318996a977eb18d367a0f04674eca0defc9"} Dec 07 09:19:51 crc kubenswrapper[4838]: I1207 09:19:51.525997 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-xrdpm" podUID="66392b25-1220-47fc-96d2-63e43a34ce70" containerName="console" containerID="cri-o://3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18" gracePeriod=15 Dec 07 09:19:51 crc kubenswrapper[4838]: I1207 09:19:51.545322 4838 generic.go:334] "Generic (PLEG): container finished" podID="36061897-0432-4e69-add8-2a3560a583b5" containerID="5e5b587525f5ec5f4fd64c1ef3795697ab55ea0e881071e2885c3d8955d42eee" exitCode=0 Dec 07 09:19:51 crc kubenswrapper[4838]: I1207 09:19:51.545382 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" event={"ID":"36061897-0432-4e69-add8-2a3560a583b5","Type":"ContainerDied","Data":"5e5b587525f5ec5f4fd64c1ef3795697ab55ea0e881071e2885c3d8955d42eee"} Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.495335 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xrdpm_66392b25-1220-47fc-96d2-63e43a34ce70/console/0.log" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.495459 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.552331 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xrdpm_66392b25-1220-47fc-96d2-63e43a34ce70/console/0.log" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.552383 4838 generic.go:334] "Generic (PLEG): container finished" podID="66392b25-1220-47fc-96d2-63e43a34ce70" containerID="3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18" exitCode=2 Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.552624 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xrdpm" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.553273 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xrdpm" event={"ID":"66392b25-1220-47fc-96d2-63e43a34ce70","Type":"ContainerDied","Data":"3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18"} Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.553304 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xrdpm" event={"ID":"66392b25-1220-47fc-96d2-63e43a34ce70","Type":"ContainerDied","Data":"9737954c9aa09a3f3ed06b38a8ea572fa244875d6f3693e94bdb97ebec3d18f9"} Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.553321 4838 scope.go:117] "RemoveContainer" containerID="3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.568340 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-oauth-config\") pod \"66392b25-1220-47fc-96d2-63e43a34ce70\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.568441 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-service-ca\") pod \"66392b25-1220-47fc-96d2-63e43a34ce70\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.568476 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-trusted-ca-bundle\") pod \"66392b25-1220-47fc-96d2-63e43a34ce70\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.568510 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66r7p\" (UniqueName: \"kubernetes.io/projected/66392b25-1220-47fc-96d2-63e43a34ce70-kube-api-access-66r7p\") pod \"66392b25-1220-47fc-96d2-63e43a34ce70\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.568593 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-console-config\") pod \"66392b25-1220-47fc-96d2-63e43a34ce70\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.568625 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-oauth-serving-cert\") pod \"66392b25-1220-47fc-96d2-63e43a34ce70\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.568707 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-serving-cert\") pod \"66392b25-1220-47fc-96d2-63e43a34ce70\" (UID: \"66392b25-1220-47fc-96d2-63e43a34ce70\") " Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.570518 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "66392b25-1220-47fc-96d2-63e43a34ce70" (UID: "66392b25-1220-47fc-96d2-63e43a34ce70"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.571221 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-service-ca" (OuterVolumeSpecName: "service-ca") pod "66392b25-1220-47fc-96d2-63e43a34ce70" (UID: "66392b25-1220-47fc-96d2-63e43a34ce70"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.571336 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-console-config" (OuterVolumeSpecName: "console-config") pod "66392b25-1220-47fc-96d2-63e43a34ce70" (UID: "66392b25-1220-47fc-96d2-63e43a34ce70"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.571527 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "66392b25-1220-47fc-96d2-63e43a34ce70" (UID: "66392b25-1220-47fc-96d2-63e43a34ce70"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.577204 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "66392b25-1220-47fc-96d2-63e43a34ce70" (UID: "66392b25-1220-47fc-96d2-63e43a34ce70"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.577774 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "66392b25-1220-47fc-96d2-63e43a34ce70" (UID: "66392b25-1220-47fc-96d2-63e43a34ce70"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.578051 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66392b25-1220-47fc-96d2-63e43a34ce70-kube-api-access-66r7p" (OuterVolumeSpecName: "kube-api-access-66r7p") pod "66392b25-1220-47fc-96d2-63e43a34ce70" (UID: "66392b25-1220-47fc-96d2-63e43a34ce70"). InnerVolumeSpecName "kube-api-access-66r7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.582070 4838 scope.go:117] "RemoveContainer" containerID="3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18" Dec 07 09:19:52 crc kubenswrapper[4838]: E1207 09:19:52.586335 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18\": container with ID starting with 3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18 not found: ID does not exist" containerID="3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.586382 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18"} err="failed to get container status \"3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18\": rpc error: code = NotFound desc = could not find container \"3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18\": container with ID starting with 3fa4469fc234caca31b94fb90c2b976e9c5f4b0961c5c640020397e4c916de18 not found: ID does not exist" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.672491 4838 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.672534 4838 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.672553 4838 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66392b25-1220-47fc-96d2-63e43a34ce70-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.672571 4838 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.672589 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66r7p\" (UniqueName: \"kubernetes.io/projected/66392b25-1220-47fc-96d2-63e43a34ce70-kube-api-access-66r7p\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.672607 4838 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.672623 4838 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66392b25-1220-47fc-96d2-63e43a34ce70-console-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.888262 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xrdpm"] Dec 07 09:19:52 crc kubenswrapper[4838]: I1207 09:19:52.893479 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-xrdpm"] Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.097681 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.178495 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-util\") pod \"36061897-0432-4e69-add8-2a3560a583b5\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.178561 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-bundle\") pod \"36061897-0432-4e69-add8-2a3560a583b5\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.179942 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-bundle" (OuterVolumeSpecName: "bundle") pod "36061897-0432-4e69-add8-2a3560a583b5" (UID: "36061897-0432-4e69-add8-2a3560a583b5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.180033 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skg2m\" (UniqueName: \"kubernetes.io/projected/36061897-0432-4e69-add8-2a3560a583b5-kube-api-access-skg2m\") pod \"36061897-0432-4e69-add8-2a3560a583b5\" (UID: \"36061897-0432-4e69-add8-2a3560a583b5\") " Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.180648 4838 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.186533 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36061897-0432-4e69-add8-2a3560a583b5-kube-api-access-skg2m" (OuterVolumeSpecName: "kube-api-access-skg2m") pod "36061897-0432-4e69-add8-2a3560a583b5" (UID: "36061897-0432-4e69-add8-2a3560a583b5"). InnerVolumeSpecName "kube-api-access-skg2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.191771 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-util" (OuterVolumeSpecName: "util") pod "36061897-0432-4e69-add8-2a3560a583b5" (UID: "36061897-0432-4e69-add8-2a3560a583b5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.282136 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skg2m\" (UniqueName: \"kubernetes.io/projected/36061897-0432-4e69-add8-2a3560a583b5-kube-api-access-skg2m\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.282198 4838 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36061897-0432-4e69-add8-2a3560a583b5-util\") on node \"crc\" DevicePath \"\"" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.563421 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" event={"ID":"36061897-0432-4e69-add8-2a3560a583b5","Type":"ContainerDied","Data":"1dc779c74d22483729c83a91505f981f3843a131f60ae0cbdfef005e0ca07ce1"} Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.564159 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dc779c74d22483729c83a91505f981f3843a131f60ae0cbdfef005e0ca07ce1" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.563456 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd" Dec 07 09:19:53 crc kubenswrapper[4838]: I1207 09:19:53.627318 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66392b25-1220-47fc-96d2-63e43a34ce70" path="/var/lib/kubelet/pods/66392b25-1220-47fc-96d2-63e43a34ce70/volumes" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.645315 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk"] Dec 07 09:20:02 crc kubenswrapper[4838]: E1207 09:20:02.646193 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36061897-0432-4e69-add8-2a3560a583b5" containerName="util" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.646207 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="36061897-0432-4e69-add8-2a3560a583b5" containerName="util" Dec 07 09:20:02 crc kubenswrapper[4838]: E1207 09:20:02.646216 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36061897-0432-4e69-add8-2a3560a583b5" containerName="pull" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.646223 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="36061897-0432-4e69-add8-2a3560a583b5" containerName="pull" Dec 07 09:20:02 crc kubenswrapper[4838]: E1207 09:20:02.646243 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36061897-0432-4e69-add8-2a3560a583b5" containerName="extract" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.646251 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="36061897-0432-4e69-add8-2a3560a583b5" containerName="extract" Dec 07 09:20:02 crc kubenswrapper[4838]: E1207 09:20:02.646263 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66392b25-1220-47fc-96d2-63e43a34ce70" containerName="console" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.646269 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="66392b25-1220-47fc-96d2-63e43a34ce70" containerName="console" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.646382 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="66392b25-1220-47fc-96d2-63e43a34ce70" containerName="console" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.646402 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="36061897-0432-4e69-add8-2a3560a583b5" containerName="extract" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.646871 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.651764 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-ngvs5" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.652131 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.652444 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.654542 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.654588 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.676946 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk"] Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.702402 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvkk8\" (UniqueName: \"kubernetes.io/projected/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-kube-api-access-wvkk8\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.702451 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-webhook-cert\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.702478 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-apiservice-cert\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.804137 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-webhook-cert\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.804348 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-apiservice-cert\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.804485 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvkk8\" (UniqueName: \"kubernetes.io/projected/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-kube-api-access-wvkk8\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.810259 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-webhook-cert\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.814918 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-apiservice-cert\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.829975 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvkk8\" (UniqueName: \"kubernetes.io/projected/53c25e7b-09e8-435d-933b-9cbbe46f5ac0-kube-api-access-wvkk8\") pod \"metallb-operator-controller-manager-7d7fb758f7-hm2jk\" (UID: \"53c25e7b-09e8-435d-933b-9cbbe46f5ac0\") " pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:02 crc kubenswrapper[4838]: I1207 09:20:02.960866 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.027684 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl"] Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.028340 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.037016 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.037274 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.038355 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-sp5lk" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.047123 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl"] Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.124370 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cda6c003-8725-4f92-a8a8-6dad25837d56-apiservice-cert\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.124664 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7vnh\" (UniqueName: \"kubernetes.io/projected/cda6c003-8725-4f92-a8a8-6dad25837d56-kube-api-access-r7vnh\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.124723 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cda6c003-8725-4f92-a8a8-6dad25837d56-webhook-cert\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.226616 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7vnh\" (UniqueName: \"kubernetes.io/projected/cda6c003-8725-4f92-a8a8-6dad25837d56-kube-api-access-r7vnh\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.226709 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cda6c003-8725-4f92-a8a8-6dad25837d56-webhook-cert\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.226743 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cda6c003-8725-4f92-a8a8-6dad25837d56-apiservice-cert\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.232467 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cda6c003-8725-4f92-a8a8-6dad25837d56-webhook-cert\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.247416 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7vnh\" (UniqueName: \"kubernetes.io/projected/cda6c003-8725-4f92-a8a8-6dad25837d56-kube-api-access-r7vnh\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.254677 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cda6c003-8725-4f92-a8a8-6dad25837d56-apiservice-cert\") pod \"metallb-operator-webhook-server-94cfbc768-26mzl\" (UID: \"cda6c003-8725-4f92-a8a8-6dad25837d56\") " pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.309554 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk"] Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.346522 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.652504 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" event={"ID":"53c25e7b-09e8-435d-933b-9cbbe46f5ac0","Type":"ContainerStarted","Data":"452113d6e41486988bf10d809b4a77a7980356d7604beb01f596075a022caab8"} Dec 07 09:20:03 crc kubenswrapper[4838]: I1207 09:20:03.801736 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl"] Dec 07 09:20:04 crc kubenswrapper[4838]: I1207 09:20:04.639716 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" event={"ID":"cda6c003-8725-4f92-a8a8-6dad25837d56","Type":"ContainerStarted","Data":"c0c58ad4bc52067cee121d4ba0ef45676b773e97236752fde785cc725f57bff1"} Dec 07 09:20:07 crc kubenswrapper[4838]: I1207 09:20:07.659943 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" event={"ID":"53c25e7b-09e8-435d-933b-9cbbe46f5ac0","Type":"ContainerStarted","Data":"76b63a9a69c62da170b60c10f2858d675a8ab32ff96306fe662cf788daa4a70f"} Dec 07 09:20:07 crc kubenswrapper[4838]: I1207 09:20:07.660308 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:07 crc kubenswrapper[4838]: I1207 09:20:07.685455 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" podStartSLOduration=2.380884771 podStartE2EDuration="5.685437455s" podCreationTimestamp="2025-12-07 09:20:02 +0000 UTC" firstStartedPulling="2025-12-07 09:20:03.318961346 +0000 UTC m=+820.026280363" lastFinishedPulling="2025-12-07 09:20:06.62351404 +0000 UTC m=+823.330833047" observedRunningTime="2025-12-07 09:20:07.6792361 +0000 UTC m=+824.386555127" watchObservedRunningTime="2025-12-07 09:20:07.685437455 +0000 UTC m=+824.392756482" Dec 07 09:20:10 crc kubenswrapper[4838]: I1207 09:20:10.680299 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" event={"ID":"cda6c003-8725-4f92-a8a8-6dad25837d56","Type":"ContainerStarted","Data":"afda896a2bed86340a8b5eb890debc1eb0f0fc860196f1068d48af3405de67df"} Dec 07 09:20:10 crc kubenswrapper[4838]: I1207 09:20:10.680898 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:10 crc kubenswrapper[4838]: I1207 09:20:10.701306 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" podStartSLOduration=1.113397856 podStartE2EDuration="7.701285306s" podCreationTimestamp="2025-12-07 09:20:03 +0000 UTC" firstStartedPulling="2025-12-07 09:20:03.816361897 +0000 UTC m=+820.523680934" lastFinishedPulling="2025-12-07 09:20:10.404249367 +0000 UTC m=+827.111568384" observedRunningTime="2025-12-07 09:20:10.697279033 +0000 UTC m=+827.404598050" watchObservedRunningTime="2025-12-07 09:20:10.701285306 +0000 UTC m=+827.408604323" Dec 07 09:20:23 crc kubenswrapper[4838]: I1207 09:20:23.351469 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-94cfbc768-26mzl" Dec 07 09:20:42 crc kubenswrapper[4838]: I1207 09:20:42.964088 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7d7fb758f7-hm2jk" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.871127 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8"] Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.872018 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.874797 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.875198 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-wwhdf" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.880903 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-x5pk6"] Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.884760 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.891231 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.891237 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.906896 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8"] Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.990988 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-vdgqx"] Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.992403 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-vdgqx" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.996014 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-4rf7n" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.996304 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.996501 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 07 09:20:43 crc kubenswrapper[4838]: I1207 09:20:43.996682 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.000713 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-rj79s"] Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.001852 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.005309 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.018769 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rj79s"] Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.054871 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-reloader\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.054911 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-metrics\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.054940 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-sockets\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.054961 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fa3fe39-d052-4d6f-881c-526b036b4bbf-cert\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.054978 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bvgc\" (UniqueName: \"kubernetes.io/projected/4fa3fe39-d052-4d6f-881c-526b036b4bbf-kube-api-access-8bvgc\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.054996 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4269\" (UniqueName: \"kubernetes.io/projected/39164f82-bb66-4fe1-a4e4-0007d9697e40-kube-api-access-t4269\") pod \"frr-k8s-webhook-server-7fcb986d4-6btm8\" (UID: \"39164f82-bb66-4fe1-a4e4-0007d9697e40\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055022 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfnhl\" (UniqueName: \"kubernetes.io/projected/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-kube-api-access-cfnhl\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055040 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39164f82-bb66-4fe1-a4e4-0007d9697e40-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6btm8\" (UID: \"39164f82-bb66-4fe1-a4e4-0007d9697e40\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055057 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055075 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metallb-excludel2\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055089 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjxpj\" (UniqueName: \"kubernetes.io/projected/1ef88d4e-7dae-4457-85a4-0c065afd26ad-kube-api-access-tjxpj\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055105 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ef88d4e-7dae-4457-85a4-0c065afd26ad-metrics-certs\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055122 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-startup\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055136 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metrics-certs\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055158 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fa3fe39-d052-4d6f-881c-526b036b4bbf-metrics-certs\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.055176 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-conf\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156214 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-reloader\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156361 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-metrics\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156389 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-sockets\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156414 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fa3fe39-d052-4d6f-881c-526b036b4bbf-cert\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156435 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bvgc\" (UniqueName: \"kubernetes.io/projected/4fa3fe39-d052-4d6f-881c-526b036b4bbf-kube-api-access-8bvgc\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156456 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4269\" (UniqueName: \"kubernetes.io/projected/39164f82-bb66-4fe1-a4e4-0007d9697e40-kube-api-access-t4269\") pod \"frr-k8s-webhook-server-7fcb986d4-6btm8\" (UID: \"39164f82-bb66-4fe1-a4e4-0007d9697e40\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156485 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfnhl\" (UniqueName: \"kubernetes.io/projected/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-kube-api-access-cfnhl\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156502 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39164f82-bb66-4fe1-a4e4-0007d9697e40-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6btm8\" (UID: \"39164f82-bb66-4fe1-a4e4-0007d9697e40\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156519 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156535 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metallb-excludel2\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156549 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjxpj\" (UniqueName: \"kubernetes.io/projected/1ef88d4e-7dae-4457-85a4-0c065afd26ad-kube-api-access-tjxpj\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156568 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ef88d4e-7dae-4457-85a4-0c065afd26ad-metrics-certs\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156584 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-startup\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156596 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metrics-certs\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156619 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fa3fe39-d052-4d6f-881c-526b036b4bbf-metrics-certs\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156636 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-conf\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156721 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-metrics\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156728 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-sockets\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.156865 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-conf\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: E1207 09:20:44.156967 4838 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 07 09:20:44 crc kubenswrapper[4838]: E1207 09:20:44.157021 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist podName:a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86 nodeName:}" failed. No retries permitted until 2025-12-07 09:20:44.657004554 +0000 UTC m=+861.364323571 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist") pod "speaker-vdgqx" (UID: "a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86") : secret "metallb-memberlist" not found Dec 07 09:20:44 crc kubenswrapper[4838]: E1207 09:20:44.157127 4838 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 07 09:20:44 crc kubenswrapper[4838]: E1207 09:20:44.157225 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metrics-certs podName:a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86 nodeName:}" failed. No retries permitted until 2025-12-07 09:20:44.65721052 +0000 UTC m=+861.364529537 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metrics-certs") pod "speaker-vdgqx" (UID: "a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86") : secret "speaker-certs-secret" not found Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.157598 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ef88d4e-7dae-4457-85a4-0c065afd26ad-reloader\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.157921 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ef88d4e-7dae-4457-85a4-0c065afd26ad-frr-startup\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.158241 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metallb-excludel2\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.162315 4838 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.165705 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fa3fe39-d052-4d6f-881c-526b036b4bbf-metrics-certs\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.167417 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39164f82-bb66-4fe1-a4e4-0007d9697e40-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6btm8\" (UID: \"39164f82-bb66-4fe1-a4e4-0007d9697e40\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.170407 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4fa3fe39-d052-4d6f-881c-526b036b4bbf-cert\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.173208 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ef88d4e-7dae-4457-85a4-0c065afd26ad-metrics-certs\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.185625 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4269\" (UniqueName: \"kubernetes.io/projected/39164f82-bb66-4fe1-a4e4-0007d9697e40-kube-api-access-t4269\") pod \"frr-k8s-webhook-server-7fcb986d4-6btm8\" (UID: \"39164f82-bb66-4fe1-a4e4-0007d9697e40\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.190487 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfnhl\" (UniqueName: \"kubernetes.io/projected/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-kube-api-access-cfnhl\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.190514 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjxpj\" (UniqueName: \"kubernetes.io/projected/1ef88d4e-7dae-4457-85a4-0c065afd26ad-kube-api-access-tjxpj\") pod \"frr-k8s-x5pk6\" (UID: \"1ef88d4e-7dae-4457-85a4-0c065afd26ad\") " pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.203072 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bvgc\" (UniqueName: \"kubernetes.io/projected/4fa3fe39-d052-4d6f-881c-526b036b4bbf-kube-api-access-8bvgc\") pod \"controller-f8648f98b-rj79s\" (UID: \"4fa3fe39-d052-4d6f-881c-526b036b4bbf\") " pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.226685 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.242738 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.368220 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.558150 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rj79s"] Dec 07 09:20:44 crc kubenswrapper[4838]: W1207 09:20:44.561225 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fa3fe39_d052_4d6f_881c_526b036b4bbf.slice/crio-e78eca25c4fdf5c20a69925419c6173c9099dbb050b4189bd5ab4e6403fa7565 WatchSource:0}: Error finding container e78eca25c4fdf5c20a69925419c6173c9099dbb050b4189bd5ab4e6403fa7565: Status 404 returned error can't find the container with id e78eca25c4fdf5c20a69925419c6173c9099dbb050b4189bd5ab4e6403fa7565 Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.654924 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8"] Dec 07 09:20:44 crc kubenswrapper[4838]: W1207 09:20:44.659261 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39164f82_bb66_4fe1_a4e4_0007d9697e40.slice/crio-a72068a572bd99a941b2c8de3150d3b88e074168652ad946d2bf0c7bdddee5ef WatchSource:0}: Error finding container a72068a572bd99a941b2c8de3150d3b88e074168652ad946d2bf0c7bdddee5ef: Status 404 returned error can't find the container with id a72068a572bd99a941b2c8de3150d3b88e074168652ad946d2bf0c7bdddee5ef Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.666199 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metrics-certs\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.666409 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: E1207 09:20:44.666561 4838 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 07 09:20:44 crc kubenswrapper[4838]: E1207 09:20:44.666624 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist podName:a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86 nodeName:}" failed. No retries permitted until 2025-12-07 09:20:45.666606849 +0000 UTC m=+862.373925866 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist") pod "speaker-vdgqx" (UID: "a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86") : secret "metallb-memberlist" not found Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.671436 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-metrics-certs\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.885630 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerStarted","Data":"305721babb058661683473e8db21a3b555100f87789ca6e7521d785a8500df98"} Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.887322 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rj79s" event={"ID":"4fa3fe39-d052-4d6f-881c-526b036b4bbf","Type":"ContainerStarted","Data":"b791e7c77c22ce0fe24e871e7a5d8b1fa47d1dc37e791835f5c0ad05926873d6"} Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.887348 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rj79s" event={"ID":"4fa3fe39-d052-4d6f-881c-526b036b4bbf","Type":"ContainerStarted","Data":"9f9eca2aacb2f1535c04810de46a24320156a44041cdf1655337a885626c0fec"} Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.887359 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rj79s" event={"ID":"4fa3fe39-d052-4d6f-881c-526b036b4bbf","Type":"ContainerStarted","Data":"e78eca25c4fdf5c20a69925419c6173c9099dbb050b4189bd5ab4e6403fa7565"} Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.887443 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.888994 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" event={"ID":"39164f82-bb66-4fe1-a4e4-0007d9697e40","Type":"ContainerStarted","Data":"a72068a572bd99a941b2c8de3150d3b88e074168652ad946d2bf0c7bdddee5ef"} Dec 07 09:20:44 crc kubenswrapper[4838]: I1207 09:20:44.909025 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-rj79s" podStartSLOduration=1.908999289 podStartE2EDuration="1.908999289s" podCreationTimestamp="2025-12-07 09:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:20:44.903696258 +0000 UTC m=+861.611015275" watchObservedRunningTime="2025-12-07 09:20:44.908999289 +0000 UTC m=+861.616318326" Dec 07 09:20:45 crc kubenswrapper[4838]: I1207 09:20:45.678367 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:45 crc kubenswrapper[4838]: I1207 09:20:45.689768 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86-memberlist\") pod \"speaker-vdgqx\" (UID: \"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86\") " pod="metallb-system/speaker-vdgqx" Dec 07 09:20:45 crc kubenswrapper[4838]: I1207 09:20:45.863633 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-vdgqx" Dec 07 09:20:45 crc kubenswrapper[4838]: W1207 09:20:45.902251 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6564a93_33f2_4ae7_9ebd_fabf8dc0ce86.slice/crio-4ea8a4b3d75c28cfeb83a078ceaac0ce3d44a67e6e62138dc67c3f796bee56ea WatchSource:0}: Error finding container 4ea8a4b3d75c28cfeb83a078ceaac0ce3d44a67e6e62138dc67c3f796bee56ea: Status 404 returned error can't find the container with id 4ea8a4b3d75c28cfeb83a078ceaac0ce3d44a67e6e62138dc67c3f796bee56ea Dec 07 09:20:46 crc kubenswrapper[4838]: I1207 09:20:46.908753 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vdgqx" event={"ID":"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86","Type":"ContainerStarted","Data":"e9696a3116b85fff04d0ef9c12d64ce7ae383cbf82da09bb32cf069148858fff"} Dec 07 09:20:46 crc kubenswrapper[4838]: I1207 09:20:46.909020 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vdgqx" event={"ID":"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86","Type":"ContainerStarted","Data":"339b836c6552899ff5813d5c527ab386d700914ce1f3bb032314cefe323063b9"} Dec 07 09:20:46 crc kubenswrapper[4838]: I1207 09:20:46.909030 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vdgqx" event={"ID":"a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86","Type":"ContainerStarted","Data":"4ea8a4b3d75c28cfeb83a078ceaac0ce3d44a67e6e62138dc67c3f796bee56ea"} Dec 07 09:20:46 crc kubenswrapper[4838]: I1207 09:20:46.909197 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-vdgqx" Dec 07 09:20:46 crc kubenswrapper[4838]: I1207 09:20:46.929479 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-vdgqx" podStartSLOduration=3.929461629 podStartE2EDuration="3.929461629s" podCreationTimestamp="2025-12-07 09:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:20:46.924071756 +0000 UTC m=+863.631390793" watchObservedRunningTime="2025-12-07 09:20:46.929461629 +0000 UTC m=+863.636780646" Dec 07 09:20:52 crc kubenswrapper[4838]: I1207 09:20:52.955030 4838 generic.go:334] "Generic (PLEG): container finished" podID="1ef88d4e-7dae-4457-85a4-0c065afd26ad" containerID="856d38b561d387b6083f68335701d59919aa80c44696615a2a5154bb62a06875" exitCode=0 Dec 07 09:20:52 crc kubenswrapper[4838]: I1207 09:20:52.955099 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerDied","Data":"856d38b561d387b6083f68335701d59919aa80c44696615a2a5154bb62a06875"} Dec 07 09:20:52 crc kubenswrapper[4838]: I1207 09:20:52.958605 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" event={"ID":"39164f82-bb66-4fe1-a4e4-0007d9697e40","Type":"ContainerStarted","Data":"734db5e817866c737b1991b4a06411cb4675ca232261a289d87bac832e901c17"} Dec 07 09:20:53 crc kubenswrapper[4838]: I1207 09:20:53.010348 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" podStartSLOduration=2.340019711 podStartE2EDuration="10.010312589s" podCreationTimestamp="2025-12-07 09:20:43 +0000 UTC" firstStartedPulling="2025-12-07 09:20:44.662707959 +0000 UTC m=+861.370026976" lastFinishedPulling="2025-12-07 09:20:52.333000837 +0000 UTC m=+869.040319854" observedRunningTime="2025-12-07 09:20:53.009757963 +0000 UTC m=+869.717077050" watchObservedRunningTime="2025-12-07 09:20:53.010312589 +0000 UTC m=+869.717631656" Dec 07 09:20:53 crc kubenswrapper[4838]: I1207 09:20:53.967537 4838 generic.go:334] "Generic (PLEG): container finished" podID="1ef88d4e-7dae-4457-85a4-0c065afd26ad" containerID="e1749e31721207d170c1d854708e71f50b2b3eb2225c80eb3b7de7443f34a14f" exitCode=0 Dec 07 09:20:53 crc kubenswrapper[4838]: I1207 09:20:53.969023 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerDied","Data":"e1749e31721207d170c1d854708e71f50b2b3eb2225c80eb3b7de7443f34a14f"} Dec 07 09:20:53 crc kubenswrapper[4838]: I1207 09:20:53.969168 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:20:54 crc kubenswrapper[4838]: I1207 09:20:54.373296 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-rj79s" Dec 07 09:20:54 crc kubenswrapper[4838]: I1207 09:20:54.978072 4838 generic.go:334] "Generic (PLEG): container finished" podID="1ef88d4e-7dae-4457-85a4-0c065afd26ad" containerID="c86131a48ded37377e8ba2adbc5996eb2ea91f4980eb79107c0e26fc2e103e0f" exitCode=0 Dec 07 09:20:54 crc kubenswrapper[4838]: I1207 09:20:54.978144 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerDied","Data":"c86131a48ded37377e8ba2adbc5996eb2ea91f4980eb79107c0e26fc2e103e0f"} Dec 07 09:20:55 crc kubenswrapper[4838]: I1207 09:20:55.987835 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerStarted","Data":"9f543e568a4f471ef3351ac9a2b1be245cf61a9200b0fdacacddeeecc4f15046"} Dec 07 09:20:56 crc kubenswrapper[4838]: I1207 09:20:56.996578 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerStarted","Data":"79e03a0c146bb3c921fab924bc5a4f906155f5e30f074a6cc4e5cfd7634576ce"} Dec 07 09:20:56 crc kubenswrapper[4838]: I1207 09:20:56.996628 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerStarted","Data":"00d0910db4a3e033cdaf65567062c6a1d390b5cdebd66cc545d60620ce4149e1"} Dec 07 09:20:56 crc kubenswrapper[4838]: I1207 09:20:56.996643 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerStarted","Data":"a37cce3c0137282b9777766df6078b70ecaa141dbd128e0929deaef66d29b44a"} Dec 07 09:20:56 crc kubenswrapper[4838]: I1207 09:20:56.996656 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerStarted","Data":"8cc3b719b3344a3080925ec0341ca23148b863b0be03ab50d8f8be176de40412"} Dec 07 09:20:56 crc kubenswrapper[4838]: I1207 09:20:56.996669 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x5pk6" event={"ID":"1ef88d4e-7dae-4457-85a4-0c065afd26ad","Type":"ContainerStarted","Data":"f22f6e5b4db81f219a3ade1552c54899454bb194ef7d8cfa726312ec11a81a98"} Dec 07 09:20:56 crc kubenswrapper[4838]: I1207 09:20:56.996717 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.019055 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-x5pk6" podStartSLOduration=6.247859185 podStartE2EDuration="14.019032642s" podCreationTimestamp="2025-12-07 09:20:43 +0000 UTC" firstStartedPulling="2025-12-07 09:20:44.547062001 +0000 UTC m=+861.254381018" lastFinishedPulling="2025-12-07 09:20:52.318235458 +0000 UTC m=+869.025554475" observedRunningTime="2025-12-07 09:20:57.017736385 +0000 UTC m=+873.725055402" watchObservedRunningTime="2025-12-07 09:20:57.019032642 +0000 UTC m=+873.726351679" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.304532 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k7sc6"] Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.306295 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.322793 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7sc6"] Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.345526 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-utilities\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.345628 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-catalog-content\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.345649 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzz2l\" (UniqueName: \"kubernetes.io/projected/350ff2de-32e0-4702-b9e6-9730841d8231-kube-api-access-xzz2l\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.448490 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-utilities\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.448888 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-catalog-content\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.448915 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzz2l\" (UniqueName: \"kubernetes.io/projected/350ff2de-32e0-4702-b9e6-9730841d8231-kube-api-access-xzz2l\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.449034 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-utilities\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.449482 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-catalog-content\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.487250 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzz2l\" (UniqueName: \"kubernetes.io/projected/350ff2de-32e0-4702-b9e6-9730841d8231-kube-api-access-xzz2l\") pod \"redhat-marketplace-k7sc6\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.619986 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:20:57 crc kubenswrapper[4838]: I1207 09:20:57.937490 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7sc6"] Dec 07 09:20:57 crc kubenswrapper[4838]: W1207 09:20:57.958954 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod350ff2de_32e0_4702_b9e6_9730841d8231.slice/crio-22d479641626b492f9cdc9d5e5fdcff76c13160c534167958de2dcc2e58278c0 WatchSource:0}: Error finding container 22d479641626b492f9cdc9d5e5fdcff76c13160c534167958de2dcc2e58278c0: Status 404 returned error can't find the container with id 22d479641626b492f9cdc9d5e5fdcff76c13160c534167958de2dcc2e58278c0 Dec 07 09:20:58 crc kubenswrapper[4838]: I1207 09:20:58.006155 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7sc6" event={"ID":"350ff2de-32e0-4702-b9e6-9730841d8231","Type":"ContainerStarted","Data":"22d479641626b492f9cdc9d5e5fdcff76c13160c534167958de2dcc2e58278c0"} Dec 07 09:20:59 crc kubenswrapper[4838]: I1207 09:20:59.013024 4838 generic.go:334] "Generic (PLEG): container finished" podID="350ff2de-32e0-4702-b9e6-9730841d8231" containerID="0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709" exitCode=0 Dec 07 09:20:59 crc kubenswrapper[4838]: I1207 09:20:59.013719 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7sc6" event={"ID":"350ff2de-32e0-4702-b9e6-9730841d8231","Type":"ContainerDied","Data":"0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709"} Dec 07 09:20:59 crc kubenswrapper[4838]: I1207 09:20:59.243157 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:20:59 crc kubenswrapper[4838]: I1207 09:20:59.298776 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:21:00 crc kubenswrapper[4838]: I1207 09:21:00.021950 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7sc6" event={"ID":"350ff2de-32e0-4702-b9e6-9730841d8231","Type":"ContainerStarted","Data":"1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442"} Dec 07 09:21:01 crc kubenswrapper[4838]: I1207 09:21:01.042291 4838 generic.go:334] "Generic (PLEG): container finished" podID="350ff2de-32e0-4702-b9e6-9730841d8231" containerID="1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442" exitCode=0 Dec 07 09:21:01 crc kubenswrapper[4838]: I1207 09:21:01.042430 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7sc6" event={"ID":"350ff2de-32e0-4702-b9e6-9730841d8231","Type":"ContainerDied","Data":"1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442"} Dec 07 09:21:02 crc kubenswrapper[4838]: I1207 09:21:02.051806 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7sc6" event={"ID":"350ff2de-32e0-4702-b9e6-9730841d8231","Type":"ContainerStarted","Data":"bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec"} Dec 07 09:21:04 crc kubenswrapper[4838]: I1207 09:21:04.233533 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6btm8" Dec 07 09:21:04 crc kubenswrapper[4838]: I1207 09:21:04.265415 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k7sc6" podStartSLOduration=4.692694015 podStartE2EDuration="7.265365031s" podCreationTimestamp="2025-12-07 09:20:57 +0000 UTC" firstStartedPulling="2025-12-07 09:20:59.015363715 +0000 UTC m=+875.722682732" lastFinishedPulling="2025-12-07 09:21:01.588034721 +0000 UTC m=+878.295353748" observedRunningTime="2025-12-07 09:21:02.078133311 +0000 UTC m=+878.785452368" watchObservedRunningTime="2025-12-07 09:21:04.265365031 +0000 UTC m=+880.972684088" Dec 07 09:21:05 crc kubenswrapper[4838]: I1207 09:21:05.871007 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-vdgqx" Dec 07 09:21:07 crc kubenswrapper[4838]: I1207 09:21:07.626610 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:21:07 crc kubenswrapper[4838]: I1207 09:21:07.626664 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:21:07 crc kubenswrapper[4838]: I1207 09:21:07.669489 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:21:08 crc kubenswrapper[4838]: I1207 09:21:08.126830 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:21:08 crc kubenswrapper[4838]: I1207 09:21:08.172189 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7sc6"] Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.100858 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k7sc6" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="registry-server" containerID="cri-o://bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec" gracePeriod=2 Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.588135 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.737850 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-utilities\") pod \"350ff2de-32e0-4702-b9e6-9730841d8231\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.737920 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-catalog-content\") pod \"350ff2de-32e0-4702-b9e6-9730841d8231\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.737968 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzz2l\" (UniqueName: \"kubernetes.io/projected/350ff2de-32e0-4702-b9e6-9730841d8231-kube-api-access-xzz2l\") pod \"350ff2de-32e0-4702-b9e6-9730841d8231\" (UID: \"350ff2de-32e0-4702-b9e6-9730841d8231\") " Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.741144 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-utilities" (OuterVolumeSpecName: "utilities") pod "350ff2de-32e0-4702-b9e6-9730841d8231" (UID: "350ff2de-32e0-4702-b9e6-9730841d8231"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.743352 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/350ff2de-32e0-4702-b9e6-9730841d8231-kube-api-access-xzz2l" (OuterVolumeSpecName: "kube-api-access-xzz2l") pod "350ff2de-32e0-4702-b9e6-9730841d8231" (UID: "350ff2de-32e0-4702-b9e6-9730841d8231"). InnerVolumeSpecName "kube-api-access-xzz2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.757402 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "350ff2de-32e0-4702-b9e6-9730841d8231" (UID: "350ff2de-32e0-4702-b9e6-9730841d8231"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.840034 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.840076 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350ff2de-32e0-4702-b9e6-9730841d8231-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:10 crc kubenswrapper[4838]: I1207 09:21:10.840092 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzz2l\" (UniqueName: \"kubernetes.io/projected/350ff2de-32e0-4702-b9e6-9730841d8231-kube-api-access-xzz2l\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.113208 4838 generic.go:334] "Generic (PLEG): container finished" podID="350ff2de-32e0-4702-b9e6-9730841d8231" containerID="bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec" exitCode=0 Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.113283 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7sc6" event={"ID":"350ff2de-32e0-4702-b9e6-9730841d8231","Type":"ContainerDied","Data":"bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec"} Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.113318 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7sc6" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.113344 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7sc6" event={"ID":"350ff2de-32e0-4702-b9e6-9730841d8231","Type":"ContainerDied","Data":"22d479641626b492f9cdc9d5e5fdcff76c13160c534167958de2dcc2e58278c0"} Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.113416 4838 scope.go:117] "RemoveContainer" containerID="bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.146763 4838 scope.go:117] "RemoveContainer" containerID="1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.176372 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7sc6"] Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.176469 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7sc6"] Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.180584 4838 scope.go:117] "RemoveContainer" containerID="0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.200117 4838 scope.go:117] "RemoveContainer" containerID="bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec" Dec 07 09:21:11 crc kubenswrapper[4838]: E1207 09:21:11.200622 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec\": container with ID starting with bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec not found: ID does not exist" containerID="bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.200682 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec"} err="failed to get container status \"bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec\": rpc error: code = NotFound desc = could not find container \"bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec\": container with ID starting with bc64507b496e4c462812c4fe04d8e903020f0200016b61bbb97db1e1d6476cec not found: ID does not exist" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.200708 4838 scope.go:117] "RemoveContainer" containerID="1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442" Dec 07 09:21:11 crc kubenswrapper[4838]: E1207 09:21:11.201067 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442\": container with ID starting with 1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442 not found: ID does not exist" containerID="1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.201111 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442"} err="failed to get container status \"1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442\": rpc error: code = NotFound desc = could not find container \"1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442\": container with ID starting with 1ffd6c5dfbc72d37439a048782dd6429d501b2de1e647cb076dec4b07a294442 not found: ID does not exist" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.201128 4838 scope.go:117] "RemoveContainer" containerID="0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709" Dec 07 09:21:11 crc kubenswrapper[4838]: E1207 09:21:11.201396 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709\": container with ID starting with 0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709 not found: ID does not exist" containerID="0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.201431 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709"} err="failed to get container status \"0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709\": rpc error: code = NotFound desc = could not find container \"0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709\": container with ID starting with 0ac1c414f85cf67ee1c5289ebca8751ea587cccd01ff053805d7ea7a14aac709 not found: ID does not exist" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.622395 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" path="/var/lib/kubelet/pods/350ff2de-32e0-4702-b9e6-9730841d8231/volumes" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.918711 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2sblj"] Dec 07 09:21:11 crc kubenswrapper[4838]: E1207 09:21:11.919081 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="extract-utilities" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.919102 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="extract-utilities" Dec 07 09:21:11 crc kubenswrapper[4838]: E1207 09:21:11.919134 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="registry-server" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.919147 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="registry-server" Dec 07 09:21:11 crc kubenswrapper[4838]: E1207 09:21:11.919165 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="extract-content" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.919178 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="extract-content" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.919366 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="350ff2de-32e0-4702-b9e6-9730841d8231" containerName="registry-server" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.920099 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.923515 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-k98z4" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.924406 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.933858 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 07 09:21:11 crc kubenswrapper[4838]: I1207 09:21:11.943005 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2sblj"] Dec 07 09:21:12 crc kubenswrapper[4838]: I1207 09:21:12.053939 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scw8b\" (UniqueName: \"kubernetes.io/projected/0851e8d2-5d7c-4e07-aed5-d743468017ec-kube-api-access-scw8b\") pod \"openstack-operator-index-2sblj\" (UID: \"0851e8d2-5d7c-4e07-aed5-d743468017ec\") " pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:12 crc kubenswrapper[4838]: I1207 09:21:12.154999 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scw8b\" (UniqueName: \"kubernetes.io/projected/0851e8d2-5d7c-4e07-aed5-d743468017ec-kube-api-access-scw8b\") pod \"openstack-operator-index-2sblj\" (UID: \"0851e8d2-5d7c-4e07-aed5-d743468017ec\") " pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:12 crc kubenswrapper[4838]: I1207 09:21:12.178498 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scw8b\" (UniqueName: \"kubernetes.io/projected/0851e8d2-5d7c-4e07-aed5-d743468017ec-kube-api-access-scw8b\") pod \"openstack-operator-index-2sblj\" (UID: \"0851e8d2-5d7c-4e07-aed5-d743468017ec\") " pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:12 crc kubenswrapper[4838]: I1207 09:21:12.246054 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:12 crc kubenswrapper[4838]: I1207 09:21:12.888511 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2sblj"] Dec 07 09:21:12 crc kubenswrapper[4838]: W1207 09:21:12.910785 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0851e8d2_5d7c_4e07_aed5_d743468017ec.slice/crio-8fcef6372f4e368cfdf9c0f9c5c1c190439bbdd1596d3e2d1e9a6a3c78270e11 WatchSource:0}: Error finding container 8fcef6372f4e368cfdf9c0f9c5c1c190439bbdd1596d3e2d1e9a6a3c78270e11: Status 404 returned error can't find the container with id 8fcef6372f4e368cfdf9c0f9c5c1c190439bbdd1596d3e2d1e9a6a3c78270e11 Dec 07 09:21:13 crc kubenswrapper[4838]: I1207 09:21:13.136541 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2sblj" event={"ID":"0851e8d2-5d7c-4e07-aed5-d743468017ec","Type":"ContainerStarted","Data":"8fcef6372f4e368cfdf9c0f9c5c1c190439bbdd1596d3e2d1e9a6a3c78270e11"} Dec 07 09:21:14 crc kubenswrapper[4838]: I1207 09:21:14.144346 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2sblj" event={"ID":"0851e8d2-5d7c-4e07-aed5-d743468017ec","Type":"ContainerStarted","Data":"a10cf10859d2d2334dd95fff0bb0376a1fe084996c3cf41265e2ae5aca447faa"} Dec 07 09:21:14 crc kubenswrapper[4838]: I1207 09:21:14.158178 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2sblj" podStartSLOduration=2.165106244 podStartE2EDuration="3.158159912s" podCreationTimestamp="2025-12-07 09:21:11 +0000 UTC" firstStartedPulling="2025-12-07 09:21:12.912639318 +0000 UTC m=+889.619958345" lastFinishedPulling="2025-12-07 09:21:13.905692996 +0000 UTC m=+890.613012013" observedRunningTime="2025-12-07 09:21:14.155514366 +0000 UTC m=+890.862833423" watchObservedRunningTime="2025-12-07 09:21:14.158159912 +0000 UTC m=+890.865478929" Dec 07 09:21:14 crc kubenswrapper[4838]: I1207 09:21:14.247032 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-x5pk6" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.128116 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r8kc4"] Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.131295 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.153449 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r8kc4"] Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.327759 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f97t\" (UniqueName: \"kubernetes.io/projected/f71c73e2-b437-4130-aa58-e2181fd791be-kube-api-access-8f97t\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.327806 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-utilities\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.327919 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-catalog-content\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.429012 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-catalog-content\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.429094 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f97t\" (UniqueName: \"kubernetes.io/projected/f71c73e2-b437-4130-aa58-e2181fd791be-kube-api-access-8f97t\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.429125 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-utilities\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.429540 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-catalog-content\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.429600 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-utilities\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.453857 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f97t\" (UniqueName: \"kubernetes.io/projected/f71c73e2-b437-4130-aa58-e2181fd791be-kube-api-access-8f97t\") pod \"community-operators-r8kc4\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.465025 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:17 crc kubenswrapper[4838]: I1207 09:21:17.989462 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r8kc4"] Dec 07 09:21:18 crc kubenswrapper[4838]: I1207 09:21:18.175617 4838 generic.go:334] "Generic (PLEG): container finished" podID="f71c73e2-b437-4130-aa58-e2181fd791be" containerID="d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477" exitCode=0 Dec 07 09:21:18 crc kubenswrapper[4838]: I1207 09:21:18.175665 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8kc4" event={"ID":"f71c73e2-b437-4130-aa58-e2181fd791be","Type":"ContainerDied","Data":"d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477"} Dec 07 09:21:18 crc kubenswrapper[4838]: I1207 09:21:18.175695 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8kc4" event={"ID":"f71c73e2-b437-4130-aa58-e2181fd791be","Type":"ContainerStarted","Data":"f82e5837027880134934f6b78708ad96745a9a9b7c33a9d6ab7e10221acbb66f"} Dec 07 09:21:19 crc kubenswrapper[4838]: I1207 09:21:19.184362 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8kc4" event={"ID":"f71c73e2-b437-4130-aa58-e2181fd791be","Type":"ContainerStarted","Data":"f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a"} Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.193254 4838 generic.go:334] "Generic (PLEG): container finished" podID="f71c73e2-b437-4130-aa58-e2181fd791be" containerID="f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a" exitCode=0 Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.193298 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8kc4" event={"ID":"f71c73e2-b437-4130-aa58-e2181fd791be","Type":"ContainerDied","Data":"f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a"} Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.717585 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-68zzc"] Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.719036 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.740567 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68zzc"] Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.786452 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-utilities\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.786496 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-catalog-content\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.786639 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lj5q\" (UniqueName: \"kubernetes.io/projected/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-kube-api-access-6lj5q\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.887335 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lj5q\" (UniqueName: \"kubernetes.io/projected/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-kube-api-access-6lj5q\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.887395 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-utilities\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.887422 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-catalog-content\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.887936 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-catalog-content\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.888023 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-utilities\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:20 crc kubenswrapper[4838]: I1207 09:21:20.910998 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lj5q\" (UniqueName: \"kubernetes.io/projected/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-kube-api-access-6lj5q\") pod \"certified-operators-68zzc\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:21 crc kubenswrapper[4838]: I1207 09:21:21.036290 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:21 crc kubenswrapper[4838]: I1207 09:21:21.209497 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8kc4" event={"ID":"f71c73e2-b437-4130-aa58-e2181fd791be","Type":"ContainerStarted","Data":"1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94"} Dec 07 09:21:21 crc kubenswrapper[4838]: I1207 09:21:21.243840 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r8kc4" podStartSLOduration=1.8499322409999999 podStartE2EDuration="4.243826004s" podCreationTimestamp="2025-12-07 09:21:17 +0000 UTC" firstStartedPulling="2025-12-07 09:21:18.176869299 +0000 UTC m=+894.884188326" lastFinishedPulling="2025-12-07 09:21:20.570763062 +0000 UTC m=+897.278082089" observedRunningTime="2025-12-07 09:21:21.238269026 +0000 UTC m=+897.945588043" watchObservedRunningTime="2025-12-07 09:21:21.243826004 +0000 UTC m=+897.951145021" Dec 07 09:21:21 crc kubenswrapper[4838]: I1207 09:21:21.329148 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68zzc"] Dec 07 09:21:22 crc kubenswrapper[4838]: I1207 09:21:22.216379 4838 generic.go:334] "Generic (PLEG): container finished" podID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerID="fafe3a80060f9fc19fea7cc3bd06b3af031a173e9c6da4fbeeef88be5839be85" exitCode=0 Dec 07 09:21:22 crc kubenswrapper[4838]: I1207 09:21:22.216462 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68zzc" event={"ID":"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e","Type":"ContainerDied","Data":"fafe3a80060f9fc19fea7cc3bd06b3af031a173e9c6da4fbeeef88be5839be85"} Dec 07 09:21:22 crc kubenswrapper[4838]: I1207 09:21:22.216727 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68zzc" event={"ID":"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e","Type":"ContainerStarted","Data":"ada23269f0c3bde99fbba69668a6d3701e769b2b741faf89ff71cf63f61badbc"} Dec 07 09:21:22 crc kubenswrapper[4838]: I1207 09:21:22.246723 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:22 crc kubenswrapper[4838]: I1207 09:21:22.246770 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:22 crc kubenswrapper[4838]: I1207 09:21:22.289425 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:23 crc kubenswrapper[4838]: I1207 09:21:23.226010 4838 generic.go:334] "Generic (PLEG): container finished" podID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerID="0dac7ff4e61dd95fd2e78db162d3135cc594ed38e1c7682481b0cb549226db74" exitCode=0 Dec 07 09:21:23 crc kubenswrapper[4838]: I1207 09:21:23.226070 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68zzc" event={"ID":"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e","Type":"ContainerDied","Data":"0dac7ff4e61dd95fd2e78db162d3135cc594ed38e1c7682481b0cb549226db74"} Dec 07 09:21:23 crc kubenswrapper[4838]: I1207 09:21:23.258478 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-2sblj" Dec 07 09:21:24 crc kubenswrapper[4838]: I1207 09:21:24.234140 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68zzc" event={"ID":"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e","Type":"ContainerStarted","Data":"01707cef5bae8b69219a01e306e2b860734ec5efcc4f6dcbc2a58235f770c575"} Dec 07 09:21:24 crc kubenswrapper[4838]: I1207 09:21:24.493558 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:21:24 crc kubenswrapper[4838]: I1207 09:21:24.493860 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.152310 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-68zzc" podStartSLOduration=4.737902199 podStartE2EDuration="6.152286941s" podCreationTimestamp="2025-12-07 09:21:20 +0000 UTC" firstStartedPulling="2025-12-07 09:21:22.21823366 +0000 UTC m=+898.925552717" lastFinishedPulling="2025-12-07 09:21:23.632618442 +0000 UTC m=+900.339937459" observedRunningTime="2025-12-07 09:21:24.273792007 +0000 UTC m=+900.981111054" watchObservedRunningTime="2025-12-07 09:21:26.152286941 +0000 UTC m=+902.859605968" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.152538 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs"] Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.154292 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.159779 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-bhrtm" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.191002 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-util\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.191299 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-bundle\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.191454 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hrgw\" (UniqueName: \"kubernetes.io/projected/3a7e3122-aebe-4d6b-8611-928f56c27d66-kube-api-access-2hrgw\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.207951 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs"] Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.291920 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-bundle\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.292043 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hrgw\" (UniqueName: \"kubernetes.io/projected/3a7e3122-aebe-4d6b-8611-928f56c27d66-kube-api-access-2hrgw\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.292089 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-util\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.292605 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-bundle\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.292673 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-util\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.316031 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hrgw\" (UniqueName: \"kubernetes.io/projected/3a7e3122-aebe-4d6b-8611-928f56c27d66-kube-api-access-2hrgw\") pod \"b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.475259 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:26 crc kubenswrapper[4838]: I1207 09:21:26.952095 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs"] Dec 07 09:21:26 crc kubenswrapper[4838]: W1207 09:21:26.964172 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a7e3122_aebe_4d6b_8611_928f56c27d66.slice/crio-c59e1b4c9a43832a325d50e57fd9227c2d17f6b3003688f9a6e3894f288581ed WatchSource:0}: Error finding container c59e1b4c9a43832a325d50e57fd9227c2d17f6b3003688f9a6e3894f288581ed: Status 404 returned error can't find the container with id c59e1b4c9a43832a325d50e57fd9227c2d17f6b3003688f9a6e3894f288581ed Dec 07 09:21:27 crc kubenswrapper[4838]: I1207 09:21:27.250672 4838 generic.go:334] "Generic (PLEG): container finished" podID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerID="238ef8db50d4b06de82597f6345b95095cbe6b760e5e3a5200bdc4a6555ec93d" exitCode=0 Dec 07 09:21:27 crc kubenswrapper[4838]: I1207 09:21:27.250726 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" event={"ID":"3a7e3122-aebe-4d6b-8611-928f56c27d66","Type":"ContainerDied","Data":"238ef8db50d4b06de82597f6345b95095cbe6b760e5e3a5200bdc4a6555ec93d"} Dec 07 09:21:27 crc kubenswrapper[4838]: I1207 09:21:27.250749 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" event={"ID":"3a7e3122-aebe-4d6b-8611-928f56c27d66","Type":"ContainerStarted","Data":"c59e1b4c9a43832a325d50e57fd9227c2d17f6b3003688f9a6e3894f288581ed"} Dec 07 09:21:27 crc kubenswrapper[4838]: I1207 09:21:27.465586 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:27 crc kubenswrapper[4838]: I1207 09:21:27.466640 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:27 crc kubenswrapper[4838]: I1207 09:21:27.529174 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:28 crc kubenswrapper[4838]: I1207 09:21:28.258789 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" event={"ID":"3a7e3122-aebe-4d6b-8611-928f56c27d66","Type":"ContainerStarted","Data":"1cd8c5a0a12dc51afcc5455d819683043b9d2267b479786b00fb3162ea383f24"} Dec 07 09:21:28 crc kubenswrapper[4838]: I1207 09:21:28.321894 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:29 crc kubenswrapper[4838]: I1207 09:21:29.267689 4838 generic.go:334] "Generic (PLEG): container finished" podID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerID="1cd8c5a0a12dc51afcc5455d819683043b9d2267b479786b00fb3162ea383f24" exitCode=0 Dec 07 09:21:29 crc kubenswrapper[4838]: I1207 09:21:29.267800 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" event={"ID":"3a7e3122-aebe-4d6b-8611-928f56c27d66","Type":"ContainerDied","Data":"1cd8c5a0a12dc51afcc5455d819683043b9d2267b479786b00fb3162ea383f24"} Dec 07 09:21:30 crc kubenswrapper[4838]: I1207 09:21:30.284983 4838 generic.go:334] "Generic (PLEG): container finished" podID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerID="e187728b856cdad5b5fd1225ca7f1a95de0f47d1f03c5c11d88d59187ec4d69e" exitCode=0 Dec 07 09:21:30 crc kubenswrapper[4838]: I1207 09:21:30.285100 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" event={"ID":"3a7e3122-aebe-4d6b-8611-928f56c27d66","Type":"ContainerDied","Data":"e187728b856cdad5b5fd1225ca7f1a95de0f47d1f03c5c11d88d59187ec4d69e"} Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.037161 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.037265 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.109007 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.347082 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.539600 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.661924 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-util\") pod \"3a7e3122-aebe-4d6b-8611-928f56c27d66\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.661983 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hrgw\" (UniqueName: \"kubernetes.io/projected/3a7e3122-aebe-4d6b-8611-928f56c27d66-kube-api-access-2hrgw\") pod \"3a7e3122-aebe-4d6b-8611-928f56c27d66\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.662134 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-bundle\") pod \"3a7e3122-aebe-4d6b-8611-928f56c27d66\" (UID: \"3a7e3122-aebe-4d6b-8611-928f56c27d66\") " Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.662716 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-bundle" (OuterVolumeSpecName: "bundle") pod "3a7e3122-aebe-4d6b-8611-928f56c27d66" (UID: "3a7e3122-aebe-4d6b-8611-928f56c27d66"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.670070 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a7e3122-aebe-4d6b-8611-928f56c27d66-kube-api-access-2hrgw" (OuterVolumeSpecName: "kube-api-access-2hrgw") pod "3a7e3122-aebe-4d6b-8611-928f56c27d66" (UID: "3a7e3122-aebe-4d6b-8611-928f56c27d66"). InnerVolumeSpecName "kube-api-access-2hrgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.686483 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-util" (OuterVolumeSpecName: "util") pod "3a7e3122-aebe-4d6b-8611-928f56c27d66" (UID: "3a7e3122-aebe-4d6b-8611-928f56c27d66"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.764342 4838 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.764378 4838 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a7e3122-aebe-4d6b-8611-928f56c27d66-util\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:31 crc kubenswrapper[4838]: I1207 09:21:31.764393 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hrgw\" (UniqueName: \"kubernetes.io/projected/3a7e3122-aebe-4d6b-8611-928f56c27d66-kube-api-access-2hrgw\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:32 crc kubenswrapper[4838]: I1207 09:21:32.303881 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" event={"ID":"3a7e3122-aebe-4d6b-8611-928f56c27d66","Type":"ContainerDied","Data":"c59e1b4c9a43832a325d50e57fd9227c2d17f6b3003688f9a6e3894f288581ed"} Dec 07 09:21:32 crc kubenswrapper[4838]: I1207 09:21:32.303936 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs" Dec 07 09:21:32 crc kubenswrapper[4838]: I1207 09:21:32.303949 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c59e1b4c9a43832a325d50e57fd9227c2d17f6b3003688f9a6e3894f288581ed" Dec 07 09:21:34 crc kubenswrapper[4838]: I1207 09:21:34.510661 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r8kc4"] Dec 07 09:21:34 crc kubenswrapper[4838]: I1207 09:21:34.511603 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r8kc4" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="registry-server" containerID="cri-o://1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94" gracePeriod=2 Dec 07 09:21:34 crc kubenswrapper[4838]: I1207 09:21:34.895131 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:34 crc kubenswrapper[4838]: I1207 09:21:34.931485 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f97t\" (UniqueName: \"kubernetes.io/projected/f71c73e2-b437-4130-aa58-e2181fd791be-kube-api-access-8f97t\") pod \"f71c73e2-b437-4130-aa58-e2181fd791be\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " Dec 07 09:21:34 crc kubenswrapper[4838]: I1207 09:21:34.954128 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71c73e2-b437-4130-aa58-e2181fd791be-kube-api-access-8f97t" (OuterVolumeSpecName: "kube-api-access-8f97t") pod "f71c73e2-b437-4130-aa58-e2181fd791be" (UID: "f71c73e2-b437-4130-aa58-e2181fd791be"). InnerVolumeSpecName "kube-api-access-8f97t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.032203 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-utilities\") pod \"f71c73e2-b437-4130-aa58-e2181fd791be\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.032252 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-catalog-content\") pod \"f71c73e2-b437-4130-aa58-e2181fd791be\" (UID: \"f71c73e2-b437-4130-aa58-e2181fd791be\") " Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.032409 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f97t\" (UniqueName: \"kubernetes.io/projected/f71c73e2-b437-4130-aa58-e2181fd791be-kube-api-access-8f97t\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.033005 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-utilities" (OuterVolumeSpecName: "utilities") pod "f71c73e2-b437-4130-aa58-e2181fd791be" (UID: "f71c73e2-b437-4130-aa58-e2181fd791be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.084777 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f71c73e2-b437-4130-aa58-e2181fd791be" (UID: "f71c73e2-b437-4130-aa58-e2181fd791be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.133707 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.133737 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f71c73e2-b437-4130-aa58-e2181fd791be-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.326615 4838 generic.go:334] "Generic (PLEG): container finished" podID="f71c73e2-b437-4130-aa58-e2181fd791be" containerID="1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94" exitCode=0 Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.326684 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8kc4" event={"ID":"f71c73e2-b437-4130-aa58-e2181fd791be","Type":"ContainerDied","Data":"1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94"} Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.326718 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8kc4" event={"ID":"f71c73e2-b437-4130-aa58-e2181fd791be","Type":"ContainerDied","Data":"f82e5837027880134934f6b78708ad96745a9a9b7c33a9d6ab7e10221acbb66f"} Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.326739 4838 scope.go:117] "RemoveContainer" containerID="1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.326946 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8kc4" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.366097 4838 scope.go:117] "RemoveContainer" containerID="f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.368184 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r8kc4"] Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.376364 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r8kc4"] Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.388620 4838 scope.go:117] "RemoveContainer" containerID="d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.407058 4838 scope.go:117] "RemoveContainer" containerID="1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94" Dec 07 09:21:35 crc kubenswrapper[4838]: E1207 09:21:35.407489 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94\": container with ID starting with 1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94 not found: ID does not exist" containerID="1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.407528 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94"} err="failed to get container status \"1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94\": rpc error: code = NotFound desc = could not find container \"1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94\": container with ID starting with 1705e9f173753b710384ee2fc56b3b51f189cbd891abb4e2551ce561d50aeb94 not found: ID does not exist" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.407553 4838 scope.go:117] "RemoveContainer" containerID="f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a" Dec 07 09:21:35 crc kubenswrapper[4838]: E1207 09:21:35.407896 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a\": container with ID starting with f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a not found: ID does not exist" containerID="f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.407940 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a"} err="failed to get container status \"f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a\": rpc error: code = NotFound desc = could not find container \"f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a\": container with ID starting with f23bf8eff9699b6b9142c8ba3b02fb06eb9e863be794e4b54c32dc77d07d001a not found: ID does not exist" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.407957 4838 scope.go:117] "RemoveContainer" containerID="d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477" Dec 07 09:21:35 crc kubenswrapper[4838]: E1207 09:21:35.408318 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477\": container with ID starting with d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477 not found: ID does not exist" containerID="d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.408351 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477"} err="failed to get container status \"d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477\": rpc error: code = NotFound desc = could not find container \"d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477\": container with ID starting with d0753b65709be82b4fc7cb01f8a6f76d0691cac0b30b217e1a8a25e293bbb477 not found: ID does not exist" Dec 07 09:21:35 crc kubenswrapper[4838]: I1207 09:21:35.621297 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" path="/var/lib/kubelet/pods/f71c73e2-b437-4130-aa58-e2181fd791be/volumes" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.100475 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68zzc"] Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.100689 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-68zzc" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="registry-server" containerID="cri-o://01707cef5bae8b69219a01e306e2b860734ec5efcc4f6dcbc2a58235f770c575" gracePeriod=2 Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.353650 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h"] Dec 07 09:21:36 crc kubenswrapper[4838]: E1207 09:21:36.353883 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerName="extract" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.353894 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerName="extract" Dec 07 09:21:36 crc kubenswrapper[4838]: E1207 09:21:36.353907 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerName="pull" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.353913 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerName="pull" Dec 07 09:21:36 crc kubenswrapper[4838]: E1207 09:21:36.353927 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="extract-utilities" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.353934 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="extract-utilities" Dec 07 09:21:36 crc kubenswrapper[4838]: E1207 09:21:36.353942 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerName="util" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.353948 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerName="util" Dec 07 09:21:36 crc kubenswrapper[4838]: E1207 09:21:36.353961 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="extract-content" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.353967 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="extract-content" Dec 07 09:21:36 crc kubenswrapper[4838]: E1207 09:21:36.353975 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="registry-server" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.353981 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="registry-server" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.354092 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a7e3122-aebe-4d6b-8611-928f56c27d66" containerName="extract" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.354104 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f71c73e2-b437-4130-aa58-e2181fd791be" containerName="registry-server" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.354465 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.356681 4838 generic.go:334] "Generic (PLEG): container finished" podID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerID="01707cef5bae8b69219a01e306e2b860734ec5efcc4f6dcbc2a58235f770c575" exitCode=0 Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.356722 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68zzc" event={"ID":"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e","Type":"ContainerDied","Data":"01707cef5bae8b69219a01e306e2b860734ec5efcc4f6dcbc2a58235f770c575"} Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.357615 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-z9475" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.413845 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h"] Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.541668 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.552247 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nmqt\" (UniqueName: \"kubernetes.io/projected/e28de98c-d75e-4673-8299-fd5abba22b44-kube-api-access-2nmqt\") pod \"openstack-operator-controller-operator-5f48db4cb9-cs44h\" (UID: \"e28de98c-d75e-4673-8299-fd5abba22b44\") " pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.653339 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-utilities\") pod \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.653390 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-catalog-content\") pod \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.653908 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lj5q\" (UniqueName: \"kubernetes.io/projected/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-kube-api-access-6lj5q\") pod \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\" (UID: \"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e\") " Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.654146 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nmqt\" (UniqueName: \"kubernetes.io/projected/e28de98c-d75e-4673-8299-fd5abba22b44-kube-api-access-2nmqt\") pod \"openstack-operator-controller-operator-5f48db4cb9-cs44h\" (UID: \"e28de98c-d75e-4673-8299-fd5abba22b44\") " pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.654172 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-utilities" (OuterVolumeSpecName: "utilities") pod "f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" (UID: "f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.668077 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-kube-api-access-6lj5q" (OuterVolumeSpecName: "kube-api-access-6lj5q") pod "f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" (UID: "f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e"). InnerVolumeSpecName "kube-api-access-6lj5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.671158 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nmqt\" (UniqueName: \"kubernetes.io/projected/e28de98c-d75e-4673-8299-fd5abba22b44-kube-api-access-2nmqt\") pod \"openstack-operator-controller-operator-5f48db4cb9-cs44h\" (UID: \"e28de98c-d75e-4673-8299-fd5abba22b44\") " pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.672721 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.699038 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" (UID: "f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.755308 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lj5q\" (UniqueName: \"kubernetes.io/projected/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-kube-api-access-6lj5q\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.755350 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:36 crc kubenswrapper[4838]: I1207 09:21:36.755364 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.223621 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h"] Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.364188 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68zzc" event={"ID":"f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e","Type":"ContainerDied","Data":"ada23269f0c3bde99fbba69668a6d3701e769b2b741faf89ff71cf63f61badbc"} Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.364233 4838 scope.go:117] "RemoveContainer" containerID="01707cef5bae8b69219a01e306e2b860734ec5efcc4f6dcbc2a58235f770c575" Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.364259 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68zzc" Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.366603 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" event={"ID":"e28de98c-d75e-4673-8299-fd5abba22b44","Type":"ContainerStarted","Data":"f52632e60547d62d16e8e801ba72daa6a2b35ee8d71ab9ca2e6fac2d23682bb1"} Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.379197 4838 scope.go:117] "RemoveContainer" containerID="0dac7ff4e61dd95fd2e78db162d3135cc594ed38e1c7682481b0cb549226db74" Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.394949 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68zzc"] Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.398753 4838 scope.go:117] "RemoveContainer" containerID="fafe3a80060f9fc19fea7cc3bd06b3af031a173e9c6da4fbeeef88be5839be85" Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.399451 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-68zzc"] Dec 07 09:21:37 crc kubenswrapper[4838]: I1207 09:21:37.623714 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" path="/var/lib/kubelet/pods/f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e/volumes" Dec 07 09:21:43 crc kubenswrapper[4838]: I1207 09:21:43.414103 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" event={"ID":"e28de98c-d75e-4673-8299-fd5abba22b44","Type":"ContainerStarted","Data":"9406191b6cdea87bde4a1c0fecf54a7acd2a379b3932927aa72395b8ee4b5035"} Dec 07 09:21:43 crc kubenswrapper[4838]: I1207 09:21:43.414650 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" Dec 07 09:21:43 crc kubenswrapper[4838]: I1207 09:21:43.454179 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" podStartSLOduration=2.358000994 podStartE2EDuration="7.454148285s" podCreationTimestamp="2025-12-07 09:21:36 +0000 UTC" firstStartedPulling="2025-12-07 09:21:37.229466336 +0000 UTC m=+913.936785353" lastFinishedPulling="2025-12-07 09:21:42.325613617 +0000 UTC m=+919.032932644" observedRunningTime="2025-12-07 09:21:43.444759118 +0000 UTC m=+920.152078165" watchObservedRunningTime="2025-12-07 09:21:43.454148285 +0000 UTC m=+920.161467342" Dec 07 09:21:54 crc kubenswrapper[4838]: I1207 09:21:54.492952 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:21:54 crc kubenswrapper[4838]: I1207 09:21:54.493527 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:21:56 crc kubenswrapper[4838]: I1207 09:21:56.679773 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5f48db4cb9-cs44h" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.108425 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx"] Dec 07 09:22:16 crc kubenswrapper[4838]: E1207 09:22:16.110630 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="extract-content" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.110739 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="extract-content" Dec 07 09:22:16 crc kubenswrapper[4838]: E1207 09:22:16.110861 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="extract-utilities" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.110968 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="extract-utilities" Dec 07 09:22:16 crc kubenswrapper[4838]: E1207 09:22:16.111057 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="registry-server" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.111136 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="registry-server" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.111355 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3bc8af4-c189-4ad8-9b5b-cc3504fe8a3e" containerName="registry-server" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.112394 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.121300 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-98vzv" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.128145 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.130486 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.133635 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-2m7gg" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.134413 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.135577 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.143640 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-g9wc4" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.143753 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.153990 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.168517 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.170203 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9qxv\" (UniqueName: \"kubernetes.io/projected/2e6976c0-74c7-4f62-88e0-6f6b635f829b-kube-api-access-h9qxv\") pod \"barbican-operator-controller-manager-7d9dfd778-5cgxx\" (UID: \"2e6976c0-74c7-4f62-88e0-6f6b635f829b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.170342 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pvcr\" (UniqueName: \"kubernetes.io/projected/c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd-kube-api-access-8pvcr\") pod \"designate-operator-controller-manager-697fb699cf-vb8wb\" (UID: \"c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.170487 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xthtc\" (UniqueName: \"kubernetes.io/projected/e9af8ce9-a7b0-4bca-b834-2f2f494eb434-kube-api-access-xthtc\") pod \"cinder-operator-controller-manager-6c677c69b-p5xdf\" (UID: \"e9af8ce9-a7b0-4bca-b834-2f2f494eb434\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.201458 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.202444 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.211570 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.212304 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-xdq8t" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.213338 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.229215 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pxhzx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.236928 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.258883 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.259861 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.266898 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.281938 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.282297 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-nmb85" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.286316 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xthtc\" (UniqueName: \"kubernetes.io/projected/e9af8ce9-a7b0-4bca-b834-2f2f494eb434-kube-api-access-xthtc\") pod \"cinder-operator-controller-manager-6c677c69b-p5xdf\" (UID: \"e9af8ce9-a7b0-4bca-b834-2f2f494eb434\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.286406 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9qxv\" (UniqueName: \"kubernetes.io/projected/2e6976c0-74c7-4f62-88e0-6f6b635f829b-kube-api-access-h9qxv\") pod \"barbican-operator-controller-manager-7d9dfd778-5cgxx\" (UID: \"2e6976c0-74c7-4f62-88e0-6f6b635f829b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.286440 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pvcr\" (UniqueName: \"kubernetes.io/projected/c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd-kube-api-access-8pvcr\") pod \"designate-operator-controller-manager-697fb699cf-vb8wb\" (UID: \"c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.313899 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.319159 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.364494 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pvcr\" (UniqueName: \"kubernetes.io/projected/c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd-kube-api-access-8pvcr\") pod \"designate-operator-controller-manager-697fb699cf-vb8wb\" (UID: \"c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.365831 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xthtc\" (UniqueName: \"kubernetes.io/projected/e9af8ce9-a7b0-4bca-b834-2f2f494eb434-kube-api-access-xthtc\") pod \"cinder-operator-controller-manager-6c677c69b-p5xdf\" (UID: \"e9af8ce9-a7b0-4bca-b834-2f2f494eb434\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.365911 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.367530 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-qll6t" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.377404 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9qxv\" (UniqueName: \"kubernetes.io/projected/2e6976c0-74c7-4f62-88e0-6f6b635f829b-kube-api-access-h9qxv\") pod \"barbican-operator-controller-manager-7d9dfd778-5cgxx\" (UID: \"2e6976c0-74c7-4f62-88e0-6f6b635f829b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.389458 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tfh6\" (UniqueName: \"kubernetes.io/projected/49a96bc4-d983-4088-973e-ebb95fdcc725-kube-api-access-5tfh6\") pod \"horizon-operator-controller-manager-68c6d99b8f-2ftqb\" (UID: \"49a96bc4-d983-4088-973e-ebb95fdcc725\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.389511 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9bt8\" (UniqueName: \"kubernetes.io/projected/f91ec7b3-50b5-4f9f-95a0-c65abeeb879f-kube-api-access-h9bt8\") pod \"glance-operator-controller-manager-5697bb5779-9hg7c\" (UID: \"f91ec7b3-50b5-4f9f-95a0-c65abeeb879f\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.389539 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z45s\" (UniqueName: \"kubernetes.io/projected/5d086115-289a-4478-a657-7ea3927165cd-kube-api-access-8z45s\") pod \"heat-operator-controller-manager-5f64f6f8bb-4hn7r\" (UID: \"5d086115-289a-4478-a657-7ea3927165cd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.397057 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.420132 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.422343 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.428792 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hv9qq" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.432435 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.433264 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.456231 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.457200 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.466263 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-mp67w" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.468638 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.486438 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.487541 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.488911 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.494672 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-bjk8b" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.495289 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9bt8\" (UniqueName: \"kubernetes.io/projected/f91ec7b3-50b5-4f9f-95a0-c65abeeb879f-kube-api-access-h9bt8\") pod \"glance-operator-controller-manager-5697bb5779-9hg7c\" (UID: \"f91ec7b3-50b5-4f9f-95a0-c65abeeb879f\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.495329 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z45s\" (UniqueName: \"kubernetes.io/projected/5d086115-289a-4478-a657-7ea3927165cd-kube-api-access-8z45s\") pod \"heat-operator-controller-manager-5f64f6f8bb-4hn7r\" (UID: \"5d086115-289a-4478-a657-7ea3927165cd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.495355 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4k56\" (UniqueName: \"kubernetes.io/projected/fac41ae4-04ea-4e07-b672-13fdf35a6530-kube-api-access-k4k56\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.495395 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.495433 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tfh6\" (UniqueName: \"kubernetes.io/projected/49a96bc4-d983-4088-973e-ebb95fdcc725-kube-api-access-5tfh6\") pod \"horizon-operator-controller-manager-68c6d99b8f-2ftqb\" (UID: \"49a96bc4-d983-4088-973e-ebb95fdcc725\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.531092 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.548378 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z45s\" (UniqueName: \"kubernetes.io/projected/5d086115-289a-4478-a657-7ea3927165cd-kube-api-access-8z45s\") pod \"heat-operator-controller-manager-5f64f6f8bb-4hn7r\" (UID: \"5d086115-289a-4478-a657-7ea3927165cd\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.562099 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.562759 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tfh6\" (UniqueName: \"kubernetes.io/projected/49a96bc4-d983-4088-973e-ebb95fdcc725-kube-api-access-5tfh6\") pod \"horizon-operator-controller-manager-68c6d99b8f-2ftqb\" (UID: \"49a96bc4-d983-4088-973e-ebb95fdcc725\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.574238 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9bt8\" (UniqueName: \"kubernetes.io/projected/f91ec7b3-50b5-4f9f-95a0-c65abeeb879f-kube-api-access-h9bt8\") pod \"glance-operator-controller-manager-5697bb5779-9hg7c\" (UID: \"f91ec7b3-50b5-4f9f-95a0-c65abeeb879f\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.597514 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pkv8\" (UniqueName: \"kubernetes.io/projected/ba02a228-91ff-41cf-88ac-b027e697dff9-kube-api-access-9pkv8\") pod \"ironic-operator-controller-manager-967d97867-pmdl9\" (UID: \"ba02a228-91ff-41cf-88ac-b027e697dff9\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.597581 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4k56\" (UniqueName: \"kubernetes.io/projected/fac41ae4-04ea-4e07-b672-13fdf35a6530-kube-api-access-k4k56\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.597636 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.597662 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64vv9\" (UniqueName: \"kubernetes.io/projected/23f05536-d749-43c5-a7cc-7dc63f46e288-kube-api-access-64vv9\") pod \"keystone-operator-controller-manager-7765d96ddf-gqrds\" (UID: \"23f05536-d749-43c5-a7cc-7dc63f46e288\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.597693 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqrgq\" (UniqueName: \"kubernetes.io/projected/7feb13b5-ed7a-47bf-82f1-b749d5b9574e-kube-api-access-sqrgq\") pod \"manila-operator-controller-manager-5b5fd79c9c-hbffh\" (UID: \"7feb13b5-ed7a-47bf-82f1-b749d5b9574e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" Dec 07 09:22:16 crc kubenswrapper[4838]: E1207 09:22:16.598564 4838 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:16 crc kubenswrapper[4838]: E1207 09:22:16.598672 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert podName:fac41ae4-04ea-4e07-b672-13fdf35a6530 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:17.098631819 +0000 UTC m=+953.805950836 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert") pod "infra-operator-controller-manager-78d48bff9d-4q7pd" (UID: "fac41ae4-04ea-4e07-b672-13fdf35a6530") : secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.602310 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.604337 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.611136 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7bvpm" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.627302 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.651678 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.653015 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.661984 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-ms8g8" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.662353 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4k56\" (UniqueName: \"kubernetes.io/projected/fac41ae4-04ea-4e07-b672-13fdf35a6530-kube-api-access-k4k56\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.666455 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.685350 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.686413 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.691223 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-b7dzh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.705635 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.706577 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64vv9\" (UniqueName: \"kubernetes.io/projected/23f05536-d749-43c5-a7cc-7dc63f46e288-kube-api-access-64vv9\") pod \"keystone-operator-controller-manager-7765d96ddf-gqrds\" (UID: \"23f05536-d749-43c5-a7cc-7dc63f46e288\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.706616 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqrgq\" (UniqueName: \"kubernetes.io/projected/7feb13b5-ed7a-47bf-82f1-b749d5b9574e-kube-api-access-sqrgq\") pod \"manila-operator-controller-manager-5b5fd79c9c-hbffh\" (UID: \"7feb13b5-ed7a-47bf-82f1-b749d5b9574e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.706644 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4zg2\" (UniqueName: \"kubernetes.io/projected/3af86462-0c95-4f05-9e30-c70da589a944-kube-api-access-z4zg2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-9cwrh\" (UID: \"3af86462-0c95-4f05-9e30-c70da589a944\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.706664 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6wvk\" (UniqueName: \"kubernetes.io/projected/3bd0361d-590b-4a06-b32c-6f7ceb427fbd-kube-api-access-b6wvk\") pod \"mariadb-operator-controller-manager-79c8c4686c-6cwcf\" (UID: \"3bd0361d-590b-4a06-b32c-6f7ceb427fbd\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.706694 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv2g9\" (UniqueName: \"kubernetes.io/projected/86196566-5b23-4cf1-924e-8a802086e443-kube-api-access-pv2g9\") pod \"nova-operator-controller-manager-697bc559fc-t6tz7\" (UID: \"86196566-5b23-4cf1-924e-8a802086e443\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.706714 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pkv8\" (UniqueName: \"kubernetes.io/projected/ba02a228-91ff-41cf-88ac-b027e697dff9-kube-api-access-9pkv8\") pod \"ironic-operator-controller-manager-967d97867-pmdl9\" (UID: \"ba02a228-91ff-41cf-88ac-b027e697dff9\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.730464 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqrgq\" (UniqueName: \"kubernetes.io/projected/7feb13b5-ed7a-47bf-82f1-b749d5b9574e-kube-api-access-sqrgq\") pod \"manila-operator-controller-manager-5b5fd79c9c-hbffh\" (UID: \"7feb13b5-ed7a-47bf-82f1-b749d5b9574e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.731584 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-54jxx"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.744652 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.748964 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.760828 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-x75tg" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.764877 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pkv8\" (UniqueName: \"kubernetes.io/projected/ba02a228-91ff-41cf-88ac-b027e697dff9-kube-api-access-9pkv8\") pod \"ironic-operator-controller-manager-967d97867-pmdl9\" (UID: \"ba02a228-91ff-41cf-88ac-b027e697dff9\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.783762 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64vv9\" (UniqueName: \"kubernetes.io/projected/23f05536-d749-43c5-a7cc-7dc63f46e288-kube-api-access-64vv9\") pod \"keystone-operator-controller-manager-7765d96ddf-gqrds\" (UID: \"23f05536-d749-43c5-a7cc-7dc63f46e288\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.787296 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-54jxx"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.795141 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.800178 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.810126 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.813037 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4zg2\" (UniqueName: \"kubernetes.io/projected/3af86462-0c95-4f05-9e30-c70da589a944-kube-api-access-z4zg2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-9cwrh\" (UID: \"3af86462-0c95-4f05-9e30-c70da589a944\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.813079 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6wvk\" (UniqueName: \"kubernetes.io/projected/3bd0361d-590b-4a06-b32c-6f7ceb427fbd-kube-api-access-b6wvk\") pod \"mariadb-operator-controller-manager-79c8c4686c-6cwcf\" (UID: \"3bd0361d-590b-4a06-b32c-6f7ceb427fbd\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.813129 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv2g9\" (UniqueName: \"kubernetes.io/projected/86196566-5b23-4cf1-924e-8a802086e443-kube-api-access-pv2g9\") pod \"nova-operator-controller-manager-697bc559fc-t6tz7\" (UID: \"86196566-5b23-4cf1-924e-8a802086e443\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.829327 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.829515 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5rbtt" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.830492 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.830934 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.844091 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.844851 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.872788 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4zg2\" (UniqueName: \"kubernetes.io/projected/3af86462-0c95-4f05-9e30-c70da589a944-kube-api-access-z4zg2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-9cwrh\" (UID: \"3af86462-0c95-4f05-9e30-c70da589a944\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.883395 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.892467 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv2g9\" (UniqueName: \"kubernetes.io/projected/86196566-5b23-4cf1-924e-8a802086e443-kube-api-access-pv2g9\") pod \"nova-operator-controller-manager-697bc559fc-t6tz7\" (UID: \"86196566-5b23-4cf1-924e-8a802086e443\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.898968 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.901215 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.906083 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.907274 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-k6lk2" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.907867 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6wvk\" (UniqueName: \"kubernetes.io/projected/3bd0361d-590b-4a06-b32c-6f7ceb427fbd-kube-api-access-b6wvk\") pod \"mariadb-operator-controller-manager-79c8c4686c-6cwcf\" (UID: \"3bd0361d-590b-4a06-b32c-6f7ceb427fbd\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.908127 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.914773 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9pd8\" (UniqueName: \"kubernetes.io/projected/07975bb2-d979-4265-bba2-1254d58b7267-kube-api-access-k9pd8\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.914887 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psrqs\" (UniqueName: \"kubernetes.io/projected/30216693-f616-448b-b85d-1c6482317ec5-kube-api-access-psrqs\") pod \"octavia-operator-controller-manager-998648c74-54jxx\" (UID: \"30216693-f616-448b-b85d-1c6482317ec5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.914912 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.916366 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.924917 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.926276 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-z4kfw" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.929362 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6"] Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.930652 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.932725 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-xrqln" Dec 07 09:22:16 crc kubenswrapper[4838]: I1207 09:22:16.933955 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.000969 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.030112 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7mws\" (UniqueName: \"kubernetes.io/projected/5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f-kube-api-access-c7mws\") pod \"swift-operator-controller-manager-9d58d64bc-6vrx6\" (UID: \"5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.030171 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psrqs\" (UniqueName: \"kubernetes.io/projected/30216693-f616-448b-b85d-1c6482317ec5-kube-api-access-psrqs\") pod \"octavia-operator-controller-manager-998648c74-54jxx\" (UID: \"30216693-f616-448b-b85d-1c6482317ec5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.030194 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.030229 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzqgq\" (UniqueName: \"kubernetes.io/projected/fdc08e45-9f41-4a0c-8478-10ab2b22ad4e-kube-api-access-pzqgq\") pod \"ovn-operator-controller-manager-b6456fdb6-b29lr\" (UID: \"fdc08e45-9f41-4a0c-8478-10ab2b22ad4e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.030251 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5sdt\" (UniqueName: \"kubernetes.io/projected/5fc12298-865b-411d-ab04-6cc9d99d8030-kube-api-access-z5sdt\") pod \"placement-operator-controller-manager-78f8948974-9jtsq\" (UID: \"5fc12298-865b-411d-ab04-6cc9d99d8030\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.030288 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9pd8\" (UniqueName: \"kubernetes.io/projected/07975bb2-d979-4265-bba2-1254d58b7267-kube-api-access-k9pd8\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.030769 4838 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.030834 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert podName:07975bb2-d979-4265-bba2-1254d58b7267 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:17.530799273 +0000 UTC m=+954.238118290 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert") pod "openstack-baremetal-operator-controller-manager-744f8cb766qxndm" (UID: "07975bb2-d979-4265-bba2-1254d58b7267") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.044328 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.068951 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.071664 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.077220 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-wt64n" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.096761 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.110829 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.129040 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psrqs\" (UniqueName: \"kubernetes.io/projected/30216693-f616-448b-b85d-1c6482317ec5-kube-api-access-psrqs\") pod \"octavia-operator-controller-manager-998648c74-54jxx\" (UID: \"30216693-f616-448b-b85d-1c6482317ec5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.134342 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.134390 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7mws\" (UniqueName: \"kubernetes.io/projected/5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f-kube-api-access-c7mws\") pod \"swift-operator-controller-manager-9d58d64bc-6vrx6\" (UID: \"5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.134459 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzqgq\" (UniqueName: \"kubernetes.io/projected/fdc08e45-9f41-4a0c-8478-10ab2b22ad4e-kube-api-access-pzqgq\") pod \"ovn-operator-controller-manager-b6456fdb6-b29lr\" (UID: \"fdc08e45-9f41-4a0c-8478-10ab2b22ad4e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.134476 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5sdt\" (UniqueName: \"kubernetes.io/projected/5fc12298-865b-411d-ab04-6cc9d99d8030-kube-api-access-z5sdt\") pod \"placement-operator-controller-manager-78f8948974-9jtsq\" (UID: \"5fc12298-865b-411d-ab04-6cc9d99d8030\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.134527 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9pd8\" (UniqueName: \"kubernetes.io/projected/07975bb2-d979-4265-bba2-1254d58b7267-kube-api-access-k9pd8\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.134807 4838 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.134873 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert podName:fac41ae4-04ea-4e07-b672-13fdf35a6530 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:18.13485758 +0000 UTC m=+954.842176597 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert") pod "infra-operator-controller-manager-78d48bff9d-4q7pd" (UID: "fac41ae4-04ea-4e07-b672-13fdf35a6530") : secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.141262 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.142750 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.145964 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-ljtrf" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.159617 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7mws\" (UniqueName: \"kubernetes.io/projected/5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f-kube-api-access-c7mws\") pod \"swift-operator-controller-manager-9d58d64bc-6vrx6\" (UID: \"5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.161499 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.165398 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzqgq\" (UniqueName: \"kubernetes.io/projected/fdc08e45-9f41-4a0c-8478-10ab2b22ad4e-kube-api-access-pzqgq\") pod \"ovn-operator-controller-manager-b6456fdb6-b29lr\" (UID: \"fdc08e45-9f41-4a0c-8478-10ab2b22ad4e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.179097 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5sdt\" (UniqueName: \"kubernetes.io/projected/5fc12298-865b-411d-ab04-6cc9d99d8030-kube-api-access-z5sdt\") pod \"placement-operator-controller-manager-78f8948974-9jtsq\" (UID: \"5fc12298-865b-411d-ab04-6cc9d99d8030\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.180002 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.181218 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.185388 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.185990 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.188040 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ctqng" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.221703 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.227627 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.228991 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.233047 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.235288 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-7rc2n" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.235522 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.235646 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.250505 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.251017 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.251045 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpl4v\" (UniqueName: \"kubernetes.io/projected/44803609-79be-4c16-bd10-3d6820d8f24b-kube-api-access-bpl4v\") pod \"test-operator-controller-manager-5854674fcc-wz9hz\" (UID: \"44803609-79be-4c16-bd10-3d6820d8f24b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.251163 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v9vm\" (UniqueName: \"kubernetes.io/projected/1f077e27-60ec-4c93-aada-4a002c07a70c-kube-api-access-4v9vm\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.251242 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-746fj\" (UniqueName: \"kubernetes.io/projected/3c038260-fe42-4c67-8e8c-b994a83c6156-kube-api-access-746fj\") pod \"watcher-operator-controller-manager-667bd8d554-v5hzl\" (UID: \"3c038260-fe42-4c67-8e8c-b994a83c6156\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.251266 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gpp7\" (UniqueName: \"kubernetes.io/projected/71940e0c-91e5-4634-92d3-d6061fbfc768-kube-api-access-2gpp7\") pod \"telemetry-operator-controller-manager-58d5ff84df-gbbg9\" (UID: \"71940e0c-91e5-4634-92d3-d6061fbfc768\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.259869 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.261463 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.267072 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-d84w8" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.277232 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.295394 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.368867 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.368914 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpl4v\" (UniqueName: \"kubernetes.io/projected/44803609-79be-4c16-bd10-3d6820d8f24b-kube-api-access-bpl4v\") pod \"test-operator-controller-manager-5854674fcc-wz9hz\" (UID: \"44803609-79be-4c16-bd10-3d6820d8f24b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.368971 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v9vm\" (UniqueName: \"kubernetes.io/projected/1f077e27-60ec-4c93-aada-4a002c07a70c-kube-api-access-4v9vm\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.369156 4838 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.369222 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:17.86920475 +0000 UTC m=+954.576523767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.369551 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-746fj\" (UniqueName: \"kubernetes.io/projected/3c038260-fe42-4c67-8e8c-b994a83c6156-kube-api-access-746fj\") pod \"watcher-operator-controller-manager-667bd8d554-v5hzl\" (UID: \"3c038260-fe42-4c67-8e8c-b994a83c6156\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.369576 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gpp7\" (UniqueName: \"kubernetes.io/projected/71940e0c-91e5-4634-92d3-d6061fbfc768-kube-api-access-2gpp7\") pod \"telemetry-operator-controller-manager-58d5ff84df-gbbg9\" (UID: \"71940e0c-91e5-4634-92d3-d6061fbfc768\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.369620 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.369695 4838 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.369736 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:17.869719885 +0000 UTC m=+954.577038902 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "metrics-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.388021 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.396902 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpl4v\" (UniqueName: \"kubernetes.io/projected/44803609-79be-4c16-bd10-3d6820d8f24b-kube-api-access-bpl4v\") pod \"test-operator-controller-manager-5854674fcc-wz9hz\" (UID: \"44803609-79be-4c16-bd10-3d6820d8f24b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.413090 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gpp7\" (UniqueName: \"kubernetes.io/projected/71940e0c-91e5-4634-92d3-d6061fbfc768-kube-api-access-2gpp7\") pod \"telemetry-operator-controller-manager-58d5ff84df-gbbg9\" (UID: \"71940e0c-91e5-4634-92d3-d6061fbfc768\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.414353 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-746fj\" (UniqueName: \"kubernetes.io/projected/3c038260-fe42-4c67-8e8c-b994a83c6156-kube-api-access-746fj\") pod \"watcher-operator-controller-manager-667bd8d554-v5hzl\" (UID: \"3c038260-fe42-4c67-8e8c-b994a83c6156\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.426109 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v9vm\" (UniqueName: \"kubernetes.io/projected/1f077e27-60ec-4c93-aada-4a002c07a70c-kube-api-access-4v9vm\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.472384 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-822vp\" (UniqueName: \"kubernetes.io/projected/700b9dd5-3c10-4fa7-bde4-0c9a4f778608-kube-api-access-822vp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tl2fn\" (UID: \"700b9dd5-3c10-4fa7-bde4-0c9a4f778608\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.562920 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.573591 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.573726 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-822vp\" (UniqueName: \"kubernetes.io/projected/700b9dd5-3c10-4fa7-bde4-0c9a4f778608-kube-api-access-822vp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tl2fn\" (UID: \"700b9dd5-3c10-4fa7-bde4-0c9a4f778608\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.574230 4838 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.574268 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert podName:07975bb2-d979-4265-bba2-1254d58b7267 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:18.574255279 +0000 UTC m=+955.281574296 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert") pod "openstack-baremetal-operator-controller-manager-744f8cb766qxndm" (UID: "07975bb2-d979-4265-bba2-1254d58b7267") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.596367 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-822vp\" (UniqueName: \"kubernetes.io/projected/700b9dd5-3c10-4fa7-bde4-0c9a4f778608-kube-api-access-822vp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tl2fn\" (UID: \"700b9dd5-3c10-4fa7-bde4-0c9a4f778608\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.611281 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.642950 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.734226 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.760188 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.805471 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx"] Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.869117 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.877275 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.877323 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.877437 4838 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.877487 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:18.877470997 +0000 UTC m=+955.584790014 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "webhook-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.877526 4838 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: E1207 09:22:17.877548 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:18.877540179 +0000 UTC m=+955.584859196 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "metrics-server-cert" not found Dec 07 09:22:17 crc kubenswrapper[4838]: I1207 09:22:17.962505 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb"] Dec 07 09:22:17 crc kubenswrapper[4838]: W1207 09:22:17.966735 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49a96bc4_d983_4088_973e_ebb95fdcc725.slice/crio-2bb14d9a18c7f92316ab4d327cf9a1b64b9d7933fc9bdc3dada6a8ceb805d7e7 WatchSource:0}: Error finding container 2bb14d9a18c7f92316ab4d327cf9a1b64b9d7933fc9bdc3dada6a8ceb805d7e7: Status 404 returned error can't find the container with id 2bb14d9a18c7f92316ab4d327cf9a1b64b9d7933fc9bdc3dada6a8ceb805d7e7 Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.181682 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.181859 4838 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.181933 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert podName:fac41ae4-04ea-4e07-b672-13fdf35a6530 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:20.181914111 +0000 UTC m=+956.889233128 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert") pod "infra-operator-controller-manager-78d48bff9d-4q7pd" (UID: "fac41ae4-04ea-4e07-b672-13fdf35a6530") : secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.367018 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.399080 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.407146 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c"] Dec 07 09:22:18 crc kubenswrapper[4838]: W1207 09:22:18.434394 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf91ec7b3_50b5_4f9f_95a0_c65abeeb879f.slice/crio-a527276d09a82944bdd3ab2864aa41b51dd2979eecbc9da98ed74bcbd8483d09 WatchSource:0}: Error finding container a527276d09a82944bdd3ab2864aa41b51dd2979eecbc9da98ed74bcbd8483d09: Status 404 returned error can't find the container with id a527276d09a82944bdd3ab2864aa41b51dd2979eecbc9da98ed74bcbd8483d09 Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.505619 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.505674 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.514790 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.556936 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6"] Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.566003 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b6wvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-6cwcf_openstack-operators(3bd0361d-590b-4a06-b32c-6f7ceb427fbd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.570972 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-746fj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-v5hzl_openstack-operators(3c038260-fe42-4c67-8e8c-b994a83c6156): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.592761 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-746fj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-v5hzl_openstack-operators(3c038260-fe42-4c67-8e8c-b994a83c6156): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.593847 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" podUID="3c038260-fe42-4c67-8e8c-b994a83c6156" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.598855 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gpp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-gbbg9_openstack-operators(71940e0c-91e5-4634-92d3-d6061fbfc768): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.599076 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pzqgq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-b29lr_openstack-operators(fdc08e45-9f41-4a0c-8478-10ab2b22ad4e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.599384 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-822vp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-tl2fn_openstack-operators(700b9dd5-3c10-4fa7-bde4-0c9a4f778608): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.600492 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" podUID="700b9dd5-3c10-4fa7-bde4-0c9a4f778608" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.603293 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.606805 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.606927 4838 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.606977 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert podName:07975bb2-d979-4265-bba2-1254d58b7267 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:20.606961272 +0000 UTC m=+957.314280279 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert") pod "openstack-baremetal-operator-controller-manager-744f8cb766qxndm" (UID: "07975bb2-d979-4265-bba2-1254d58b7267") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.608436 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-psrqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-54jxx_openstack-operators(30216693-f616-448b-b85d-1c6482317ec5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.608478 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gpp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-gbbg9_openstack-operators(71940e0c-91e5-4634-92d3-d6061fbfc768): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.608591 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pzqgq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-b29lr_openstack-operators(fdc08e45-9f41-4a0c-8478-10ab2b22ad4e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.610195 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" podUID="fdc08e45-9f41-4a0c-8478-10ab2b22ad4e" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.610309 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" podUID="71940e0c-91e5-4634-92d3-d6061fbfc768" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.610474 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bpl4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wz9hz_openstack-operators(44803609-79be-4c16-bd10-3d6820d8f24b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.610680 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-psrqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-54jxx_openstack-operators(30216693-f616-448b-b85d-1c6482317ec5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.612016 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" podUID="30216693-f616-448b-b85d-1c6482317ec5" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.612261 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9"] Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.613060 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bpl4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wz9hz_openstack-operators(44803609-79be-4c16-bd10-3d6820d8f24b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.614217 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" podUID="44803609-79be-4c16-bd10-3d6820d8f24b" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.616527 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.621172 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.626854 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.631613 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-54jxx"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.636481 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.641639 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn"] Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.642339 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" event={"ID":"f91ec7b3-50b5-4f9f-95a0-c65abeeb879f","Type":"ContainerStarted","Data":"a527276d09a82944bdd3ab2864aa41b51dd2979eecbc9da98ed74bcbd8483d09"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.646535 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" event={"ID":"fdc08e45-9f41-4a0c-8478-10ab2b22ad4e","Type":"ContainerStarted","Data":"40bd836e931aa7b627942f75a914bb683f9ac05a40990a327a2ebf3ed921ba35"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.648492 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz"] Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.651042 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" podUID="fdc08e45-9f41-4a0c-8478-10ab2b22ad4e" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.652723 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" event={"ID":"5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f","Type":"ContainerStarted","Data":"b258fb9eee69883a204d5afdd427602d8882b5fe81566060aeaeb404b69e19fb"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.655007 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" event={"ID":"2e6976c0-74c7-4f62-88e0-6f6b635f829b","Type":"ContainerStarted","Data":"1c3046c408cc8683830de300989552f7ebb379db16d2358f6d4740ebf92e7e3b"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.657844 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" event={"ID":"7feb13b5-ed7a-47bf-82f1-b749d5b9574e","Type":"ContainerStarted","Data":"610e040c481efaccab733f4b545344ac76cc0263355d73bfea89c446ebf5acd7"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.664112 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" event={"ID":"49a96bc4-d983-4088-973e-ebb95fdcc725","Type":"ContainerStarted","Data":"2bb14d9a18c7f92316ab4d327cf9a1b64b9d7933fc9bdc3dada6a8ceb805d7e7"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.673589 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" event={"ID":"e9af8ce9-a7b0-4bca-b834-2f2f494eb434","Type":"ContainerStarted","Data":"8815173fea4ff36e20eeb37a5d46a11d627687d73c7e946a62642292da7b1e27"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.675570 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" event={"ID":"3bd0361d-590b-4a06-b32c-6f7ceb427fbd","Type":"ContainerStarted","Data":"4ba30e633fea723676bd99c4141cf7c86a58c626153ffc88c29833512cdaaf9d"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.677180 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" event={"ID":"c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd","Type":"ContainerStarted","Data":"4e69e6d26deca8eaa729af242718a5bc88c481258137415a03a08317a821ec51"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.679652 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" event={"ID":"30216693-f616-448b-b85d-1c6482317ec5","Type":"ContainerStarted","Data":"729fb2e5f887df2b717a4fb3453f388de3e0c9a9b3070836318c74abfd655561"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.681948 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" event={"ID":"44803609-79be-4c16-bd10-3d6820d8f24b","Type":"ContainerStarted","Data":"f1a1b4db024e1b16704aa294f4f31b4a524cc81a27e83bb7cd15ba9b1e285ad8"} Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.684349 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" podUID="30216693-f616-448b-b85d-1c6482317ec5" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.684558 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" podUID="44803609-79be-4c16-bd10-3d6820d8f24b" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.686924 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" event={"ID":"86196566-5b23-4cf1-924e-8a802086e443","Type":"ContainerStarted","Data":"5032248bbace01398c6ec748c30625a18ece5a46c945ad4c8286fb4ba35563c7"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.691185 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" event={"ID":"23f05536-d749-43c5-a7cc-7dc63f46e288","Type":"ContainerStarted","Data":"4cdf239758d83765dbba45d3da506abdf299d8dbc47b9d3f8b9101e9359df247"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.694960 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" event={"ID":"71940e0c-91e5-4634-92d3-d6061fbfc768","Type":"ContainerStarted","Data":"87f35a867225150f022d43a465f3209f9f37837b64b7b014529c4a37428a07a8"} Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.700074 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" podUID="71940e0c-91e5-4634-92d3-d6061fbfc768" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.700546 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" event={"ID":"5fc12298-865b-411d-ab04-6cc9d99d8030","Type":"ContainerStarted","Data":"784ab2b4d6d1e264feed5e9a3fd28f4fbfac526c05ef0b318ecc0e2f93fd42c2"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.704154 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" event={"ID":"3c038260-fe42-4c67-8e8c-b994a83c6156","Type":"ContainerStarted","Data":"35a1fbbaedc0b0b34817ffff954e802cc7e9dcd59e9bc08a83ffa038306b5064"} Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.706057 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" podUID="3c038260-fe42-4c67-8e8c-b994a83c6156" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.707772 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" event={"ID":"700b9dd5-3c10-4fa7-bde4-0c9a4f778608","Type":"ContainerStarted","Data":"240f04d55ec0a4b31edc1807633b42587b0b07d26c0d5e4b95c4a96d51bcbd81"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.708888 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" event={"ID":"3af86462-0c95-4f05-9e30-c70da589a944","Type":"ContainerStarted","Data":"cd039faeb1e548bfdbe60b032ae57561f7ffe690909c43951734c75381b50af2"} Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.714294 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" podUID="700b9dd5-3c10-4fa7-bde4-0c9a4f778608" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.715117 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" event={"ID":"5d086115-289a-4478-a657-7ea3927165cd","Type":"ContainerStarted","Data":"b31a673eca72745638bc8bd8a71df1c393e54a0a8da0e8387d27c505083a7b75"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.717572 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" event={"ID":"ba02a228-91ff-41cf-88ac-b027e697dff9","Type":"ContainerStarted","Data":"5c75f06e174ba6d6e0836595171dceb44efcd15369c656525bbe0aacea4c7479"} Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.916391 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:18 crc kubenswrapper[4838]: I1207 09:22:18.916508 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.916591 4838 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.916667 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:20.916649245 +0000 UTC m=+957.623968262 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "metrics-server-cert" not found Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.916870 4838 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 09:22:18 crc kubenswrapper[4838]: E1207 09:22:18.916943 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:20.916926203 +0000 UTC m=+957.624245220 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "webhook-server-cert" not found Dec 07 09:22:19 crc kubenswrapper[4838]: E1207 09:22:19.750508 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" podUID="44803609-79be-4c16-bd10-3d6820d8f24b" Dec 07 09:22:19 crc kubenswrapper[4838]: E1207 09:22:19.750925 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" podUID="700b9dd5-3c10-4fa7-bde4-0c9a4f778608" Dec 07 09:22:19 crc kubenswrapper[4838]: E1207 09:22:19.750974 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" podUID="30216693-f616-448b-b85d-1c6482317ec5" Dec 07 09:22:19 crc kubenswrapper[4838]: E1207 09:22:19.751010 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" podUID="71940e0c-91e5-4634-92d3-d6061fbfc768" Dec 07 09:22:19 crc kubenswrapper[4838]: E1207 09:22:19.752332 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" podUID="3c038260-fe42-4c67-8e8c-b994a83c6156" Dec 07 09:22:19 crc kubenswrapper[4838]: E1207 09:22:19.773862 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" podUID="fdc08e45-9f41-4a0c-8478-10ab2b22ad4e" Dec 07 09:22:20 crc kubenswrapper[4838]: I1207 09:22:20.236291 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.236478 4838 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.236549 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert podName:fac41ae4-04ea-4e07-b672-13fdf35a6530 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:24.236532001 +0000 UTC m=+960.943851008 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert") pod "infra-operator-controller-manager-78d48bff9d-4q7pd" (UID: "fac41ae4-04ea-4e07-b672-13fdf35a6530") : secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:20 crc kubenswrapper[4838]: I1207 09:22:20.642738 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.642901 4838 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.643066 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert podName:07975bb2-d979-4265-bba2-1254d58b7267 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:24.643045996 +0000 UTC m=+961.350365013 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert") pod "openstack-baremetal-operator-controller-manager-744f8cb766qxndm" (UID: "07975bb2-d979-4265-bba2-1254d58b7267") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:20 crc kubenswrapper[4838]: I1207 09:22:20.945771 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:20 crc kubenswrapper[4838]: I1207 09:22:20.945842 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.945924 4838 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.945914 4838 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.945969 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:24.945955195 +0000 UTC m=+961.653274212 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "webhook-server-cert" not found Dec 07 09:22:20 crc kubenswrapper[4838]: E1207 09:22:20.946039 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:24.946026347 +0000 UTC m=+961.653345354 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "metrics-server-cert" not found Dec 07 09:22:24 crc kubenswrapper[4838]: I1207 09:22:24.290509 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:24 crc kubenswrapper[4838]: E1207 09:22:24.290731 4838 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:24 crc kubenswrapper[4838]: E1207 09:22:24.291108 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert podName:fac41ae4-04ea-4e07-b672-13fdf35a6530 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:32.291078897 +0000 UTC m=+968.998397964 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert") pod "infra-operator-controller-manager-78d48bff9d-4q7pd" (UID: "fac41ae4-04ea-4e07-b672-13fdf35a6530") : secret "infra-operator-webhook-server-cert" not found Dec 07 09:22:24 crc kubenswrapper[4838]: I1207 09:22:24.492994 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:22:24 crc kubenswrapper[4838]: I1207 09:22:24.493465 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:22:24 crc kubenswrapper[4838]: I1207 09:22:24.493768 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:22:24 crc kubenswrapper[4838]: I1207 09:22:24.494784 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b20dfe6b2c79e165940e9237c72516b3c8e4bccb48b171f998bf1346e1ad6571"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:22:24 crc kubenswrapper[4838]: I1207 09:22:24.495099 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://b20dfe6b2c79e165940e9237c72516b3c8e4bccb48b171f998bf1346e1ad6571" gracePeriod=600 Dec 07 09:22:24 crc kubenswrapper[4838]: I1207 09:22:24.697590 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:24 crc kubenswrapper[4838]: E1207 09:22:24.697793 4838 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:24 crc kubenswrapper[4838]: E1207 09:22:24.697909 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert podName:07975bb2-d979-4265-bba2-1254d58b7267 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:32.697885098 +0000 UTC m=+969.405204145 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert") pod "openstack-baremetal-operator-controller-manager-744f8cb766qxndm" (UID: "07975bb2-d979-4265-bba2-1254d58b7267") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:25 crc kubenswrapper[4838]: I1207 09:22:25.003090 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:25 crc kubenswrapper[4838]: E1207 09:22:25.003294 4838 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 09:22:25 crc kubenswrapper[4838]: E1207 09:22:25.003579 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:33.003553737 +0000 UTC m=+969.710872754 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "metrics-server-cert" not found Dec 07 09:22:25 crc kubenswrapper[4838]: I1207 09:22:25.003662 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:25 crc kubenswrapper[4838]: E1207 09:22:25.003690 4838 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 09:22:25 crc kubenswrapper[4838]: E1207 09:22:25.003883 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:33.003865816 +0000 UTC m=+969.711184833 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "webhook-server-cert" not found Dec 07 09:22:25 crc kubenswrapper[4838]: I1207 09:22:25.947929 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="b20dfe6b2c79e165940e9237c72516b3c8e4bccb48b171f998bf1346e1ad6571" exitCode=0 Dec 07 09:22:25 crc kubenswrapper[4838]: I1207 09:22:25.947983 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"b20dfe6b2c79e165940e9237c72516b3c8e4bccb48b171f998bf1346e1ad6571"} Dec 07 09:22:25 crc kubenswrapper[4838]: I1207 09:22:25.948021 4838 scope.go:117] "RemoveContainer" containerID="4e006964b42caddec4f66d1cf7932894d0547718b0bf611546057b7159422840" Dec 07 09:22:30 crc kubenswrapper[4838]: E1207 09:22:30.344234 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 07 09:22:30 crc kubenswrapper[4838]: E1207 09:22:30.345120 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h9bt8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-9hg7c_openstack-operators(f91ec7b3-50b5-4f9f-95a0-c65abeeb879f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:32 crc kubenswrapper[4838]: I1207 09:22:32.320545 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:32 crc kubenswrapper[4838]: I1207 09:22:32.331589 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac41ae4-04ea-4e07-b672-13fdf35a6530-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4q7pd\" (UID: \"fac41ae4-04ea-4e07-b672-13fdf35a6530\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:32 crc kubenswrapper[4838]: I1207 09:22:32.616593 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-qll6t" Dec 07 09:22:32 crc kubenswrapper[4838]: I1207 09:22:32.624362 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:32 crc kubenswrapper[4838]: I1207 09:22:32.726440 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:32 crc kubenswrapper[4838]: E1207 09:22:32.726600 4838 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:32 crc kubenswrapper[4838]: E1207 09:22:32.726653 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert podName:07975bb2-d979-4265-bba2-1254d58b7267 nodeName:}" failed. No retries permitted until 2025-12-07 09:22:48.726636887 +0000 UTC m=+985.433955914 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert") pod "openstack-baremetal-operator-controller-manager-744f8cb766qxndm" (UID: "07975bb2-d979-4265-bba2-1254d58b7267") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 09:22:32 crc kubenswrapper[4838]: E1207 09:22:32.896485 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 07 09:22:32 crc kubenswrapper[4838]: E1207 09:22:32.896719 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z4zg2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-9cwrh_openstack-operators(3af86462-0c95-4f05-9e30-c70da589a944): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:33 crc kubenswrapper[4838]: I1207 09:22:33.033225 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:33 crc kubenswrapper[4838]: I1207 09:22:33.033388 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:33 crc kubenswrapper[4838]: E1207 09:22:33.033399 4838 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 09:22:33 crc kubenswrapper[4838]: E1207 09:22:33.033507 4838 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 09:22:33 crc kubenswrapper[4838]: E1207 09:22:33.033660 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:49.033636863 +0000 UTC m=+985.740955900 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "metrics-server-cert" not found Dec 07 09:22:33 crc kubenswrapper[4838]: E1207 09:22:33.033750 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs podName:1f077e27-60ec-4c93-aada-4a002c07a70c nodeName:}" failed. No retries permitted until 2025-12-07 09:22:49.033676044 +0000 UTC m=+985.740995071 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs") pod "openstack-operator-controller-manager-5b47684954-7mp9q" (UID: "1f077e27-60ec-4c93-aada-4a002c07a70c") : secret "webhook-server-cert" not found Dec 07 09:22:37 crc kubenswrapper[4838]: E1207 09:22:37.503191 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 07 09:22:37 crc kubenswrapper[4838]: E1207 09:22:37.503616 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5tfh6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-2ftqb_openstack-operators(49a96bc4-d983-4088-973e-ebb95fdcc725): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:37 crc kubenswrapper[4838]: E1207 09:22:37.998320 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 07 09:22:37 crc kubenswrapper[4838]: E1207 09:22:37.998523 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9pkv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-pmdl9_openstack-operators(ba02a228-91ff-41cf-88ac-b027e697dff9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:39 crc kubenswrapper[4838]: E1207 09:22:39.916875 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 07 09:22:39 crc kubenswrapper[4838]: E1207 09:22:39.917398 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z5sdt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-9jtsq_openstack-operators(5fc12298-865b-411d-ab04-6cc9d99d8030): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:40 crc kubenswrapper[4838]: E1207 09:22:40.389150 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 07 09:22:40 crc kubenswrapper[4838]: E1207 09:22:40.389303 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c7mws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-6vrx6_openstack-operators(5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:41 crc kubenswrapper[4838]: E1207 09:22:41.152443 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 07 09:22:41 crc kubenswrapper[4838]: E1207 09:22:41.152672 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-64vv9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-gqrds_openstack-operators(23f05536-d749-43c5-a7cc-7dc63f46e288): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:44 crc kubenswrapper[4838]: E1207 09:22:44.837463 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8" Dec 07 09:22:44 crc kubenswrapper[4838]: E1207 09:22:44.837647 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-746fj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-v5hzl_openstack-operators(3c038260-fe42-4c67-8e8c-b994a83c6156): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:45 crc kubenswrapper[4838]: E1207 09:22:45.319604 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 07 09:22:45 crc kubenswrapper[4838]: E1207 09:22:45.319771 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pv2g9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-t6tz7_openstack-operators(86196566-5b23-4cf1-924e-8a802086e443): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:48 crc kubenswrapper[4838]: I1207 09:22:48.766652 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:48 crc kubenswrapper[4838]: I1207 09:22:48.773769 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07975bb2-d979-4265-bba2-1254d58b7267-cert\") pod \"openstack-baremetal-operator-controller-manager-744f8cb766qxndm\" (UID: \"07975bb2-d979-4265-bba2-1254d58b7267\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.032910 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5rbtt" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.041737 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.073645 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.073729 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.079901 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-webhook-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.092467 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f077e27-60ec-4c93-aada-4a002c07a70c-metrics-certs\") pod \"openstack-operator-controller-manager-5b47684954-7mp9q\" (UID: \"1f077e27-60ec-4c93-aada-4a002c07a70c\") " pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.351955 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-7rc2n" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.357036 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:49 crc kubenswrapper[4838]: I1207 09:22:49.950550 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd"] Dec 07 09:22:51 crc kubenswrapper[4838]: W1207 09:22:51.776700 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfac41ae4_04ea_4e07_b672_13fdf35a6530.slice/crio-bbbb8edc6a1ff27f36df0565c53c5c3900cb18ff47cd0cdc5c29d42834db99ce WatchSource:0}: Error finding container bbbb8edc6a1ff27f36df0565c53c5c3900cb18ff47cd0cdc5c29d42834db99ce: Status 404 returned error can't find the container with id bbbb8edc6a1ff27f36df0565c53c5c3900cb18ff47cd0cdc5c29d42834db99ce Dec 07 09:22:52 crc kubenswrapper[4838]: I1207 09:22:52.123546 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" event={"ID":"fac41ae4-04ea-4e07-b672-13fdf35a6530","Type":"ContainerStarted","Data":"bbbb8edc6a1ff27f36df0565c53c5c3900cb18ff47cd0cdc5c29d42834db99ce"} Dec 07 09:22:52 crc kubenswrapper[4838]: E1207 09:22:52.258179 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 07 09:22:52 crc kubenswrapper[4838]: E1207 09:22:52.258474 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-822vp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-tl2fn_openstack-operators(700b9dd5-3c10-4fa7-bde4-0c9a4f778608): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:52 crc kubenswrapper[4838]: E1207 09:22:52.259740 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" podUID="700b9dd5-3c10-4fa7-bde4-0c9a4f778608" Dec 07 09:22:52 crc kubenswrapper[4838]: E1207 09:22:52.832433 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 07 09:22:52 crc kubenswrapper[4838]: E1207 09:22:52.832683 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b6wvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-6cwcf_openstack-operators(3bd0361d-590b-4a06-b32c-6f7ceb427fbd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:22:52 crc kubenswrapper[4838]: E1207 09:22:52.834407 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" podUID="3bd0361d-590b-4a06-b32c-6f7ceb427fbd" Dec 07 09:22:53 crc kubenswrapper[4838]: I1207 09:22:53.168717 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"ed0d936715e5d0425ff8433d1f07f45e763ba568568132ddb5adcdb73cc5a936"} Dec 07 09:22:53 crc kubenswrapper[4838]: I1207 09:22:53.379120 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm"] Dec 07 09:22:53 crc kubenswrapper[4838]: W1207 09:22:53.390069 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07975bb2_d979_4265_bba2_1254d58b7267.slice/crio-957216f114689beecb3d5a118c2263d9c836d39b9790b9b61bd5170c0cc54ff7 WatchSource:0}: Error finding container 957216f114689beecb3d5a118c2263d9c836d39b9790b9b61bd5170c0cc54ff7: Status 404 returned error can't find the container with id 957216f114689beecb3d5a118c2263d9c836d39b9790b9b61bd5170c0cc54ff7 Dec 07 09:22:53 crc kubenswrapper[4838]: I1207 09:22:53.523615 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q"] Dec 07 09:22:53 crc kubenswrapper[4838]: W1207 09:22:53.530824 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f077e27_60ec_4c93_aada_4a002c07a70c.slice/crio-8252a97e7db1e533ac79a3d99ec1bef1fb21d45bb86fb62509143aa02474887d WatchSource:0}: Error finding container 8252a97e7db1e533ac79a3d99ec1bef1fb21d45bb86fb62509143aa02474887d: Status 404 returned error can't find the container with id 8252a97e7db1e533ac79a3d99ec1bef1fb21d45bb86fb62509143aa02474887d Dec 07 09:22:54 crc kubenswrapper[4838]: I1207 09:22:54.173236 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" event={"ID":"7feb13b5-ed7a-47bf-82f1-b749d5b9574e","Type":"ContainerStarted","Data":"123edeeccc9cbba74666cdaf107a2a1aac3afa885d23d24ca4726868777620c6"} Dec 07 09:22:54 crc kubenswrapper[4838]: I1207 09:22:54.174087 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" event={"ID":"07975bb2-d979-4265-bba2-1254d58b7267","Type":"ContainerStarted","Data":"957216f114689beecb3d5a118c2263d9c836d39b9790b9b61bd5170c0cc54ff7"} Dec 07 09:22:54 crc kubenswrapper[4838]: I1207 09:22:54.175112 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" event={"ID":"2e6976c0-74c7-4f62-88e0-6f6b635f829b","Type":"ContainerStarted","Data":"a5094155d7cc1bea2f1c9378d23282f151daba7acc6a3f4322dbde1861b74a1a"} Dec 07 09:22:54 crc kubenswrapper[4838]: I1207 09:22:54.175835 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" event={"ID":"1f077e27-60ec-4c93-aada-4a002c07a70c","Type":"ContainerStarted","Data":"8252a97e7db1e533ac79a3d99ec1bef1fb21d45bb86fb62509143aa02474887d"} Dec 07 09:22:54 crc kubenswrapper[4838]: I1207 09:22:54.501120 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:22:55 crc kubenswrapper[4838]: I1207 09:22:55.182062 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" event={"ID":"e9af8ce9-a7b0-4bca-b834-2f2f494eb434","Type":"ContainerStarted","Data":"52e64c2f07a16a296c028307dee0f3f9d52b28b2f405d4fc621d5234a9c98c25"} Dec 07 09:22:55 crc kubenswrapper[4838]: I1207 09:22:55.184173 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" event={"ID":"c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd","Type":"ContainerStarted","Data":"e0f3ba7ddba42c329582ab1c49f2de8d215f7a8b7c6166f4dce2bb34771760e7"} Dec 07 09:22:55 crc kubenswrapper[4838]: I1207 09:22:55.185627 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" event={"ID":"5d086115-289a-4478-a657-7ea3927165cd","Type":"ContainerStarted","Data":"dc6e210f4a1200e085fdfc8575d2715b495d4f8fa0dff48ab979c1de1d90e775"} Dec 07 09:22:55 crc kubenswrapper[4838]: I1207 09:22:55.187430 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" event={"ID":"71940e0c-91e5-4634-92d3-d6061fbfc768","Type":"ContainerStarted","Data":"18ef31301c18010425c1cd936774b1f0e2df5967df2395a88d8653ee2e861d58"} Dec 07 09:22:57 crc kubenswrapper[4838]: E1207 09:22:57.049498 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" podUID="23f05536-d749-43c5-a7cc-7dc63f46e288" Dec 07 09:22:57 crc kubenswrapper[4838]: E1207 09:22:57.203944 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" podUID="f91ec7b3-50b5-4f9f-95a0-c65abeeb879f" Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.204326 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" event={"ID":"23f05536-d749-43c5-a7cc-7dc63f46e288","Type":"ContainerStarted","Data":"5c5ef1757b3a6221f689039ea91236b8a56b232c5065dbeb56d964d9e438e678"} Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.214577 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" event={"ID":"30216693-f616-448b-b85d-1c6482317ec5","Type":"ContainerStarted","Data":"9e4f80619547e93fd3ff3163e25a6019bc211f93f2e89a9a4151b6aebea71c12"} Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.273930 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" event={"ID":"2e6976c0-74c7-4f62-88e0-6f6b635f829b","Type":"ContainerStarted","Data":"8e6b84fe25bed0ed96fd0c97405e2076b7b5d49271d1b18e5bfe46e60957e168"} Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.274025 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.288547 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" event={"ID":"1f077e27-60ec-4c93-aada-4a002c07a70c","Type":"ContainerStarted","Data":"e3e9265f3552e2744b8d4a3fae3ef49a1baaffd31312fa5d31b905fa170ba7b3"} Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.289571 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.295413 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" event={"ID":"44803609-79be-4c16-bd10-3d6820d8f24b","Type":"ContainerStarted","Data":"06bafa106ecb10860fb785663155927dd7e1ea6751ff4e7b45475618cd49065e"} Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.307307 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" podStartSLOduration=13.869211759 podStartE2EDuration="41.307279297s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:17.876239562 +0000 UTC m=+954.583558579" lastFinishedPulling="2025-12-07 09:22:45.3143071 +0000 UTC m=+982.021626117" observedRunningTime="2025-12-07 09:22:57.296687585 +0000 UTC m=+994.004006612" watchObservedRunningTime="2025-12-07 09:22:57.307279297 +0000 UTC m=+994.014598314" Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.351332 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" event={"ID":"f91ec7b3-50b5-4f9f-95a0-c65abeeb879f","Type":"ContainerStarted","Data":"1a6ab26215ed30112192002e2c7864be45e114673ba5c3ec18296b5a87b81473"} Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.376352 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" podStartSLOduration=40.376327023 podStartE2EDuration="40.376327023s" podCreationTimestamp="2025-12-07 09:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:22:57.370864047 +0000 UTC m=+994.078183094" watchObservedRunningTime="2025-12-07 09:22:57.376327023 +0000 UTC m=+994.083646040" Dec 07 09:22:57 crc kubenswrapper[4838]: I1207 09:22:57.376577 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" event={"ID":"fdc08e45-9f41-4a0c-8478-10ab2b22ad4e","Type":"ContainerStarted","Data":"77647ab097a7a64a9e291ba804ad55f6fc7f4a08588f3b54b354899a56278595"} Dec 07 09:22:57 crc kubenswrapper[4838]: E1207 09:22:57.512121 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" podUID="3af86462-0c95-4f05-9e30-c70da589a944" Dec 07 09:22:57 crc kubenswrapper[4838]: E1207 09:22:57.942482 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" podUID="ba02a228-91ff-41cf-88ac-b027e697dff9" Dec 07 09:22:58 crc kubenswrapper[4838]: E1207 09:22:58.119761 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" podUID="3c038260-fe42-4c67-8e8c-b994a83c6156" Dec 07 09:22:58 crc kubenswrapper[4838]: E1207 09:22:58.153979 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" podUID="86196566-5b23-4cf1-924e-8a802086e443" Dec 07 09:22:58 crc kubenswrapper[4838]: E1207 09:22:58.195123 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" podUID="5fc12298-865b-411d-ab04-6cc9d99d8030" Dec 07 09:22:58 crc kubenswrapper[4838]: E1207 09:22:58.210100 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" podUID="5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f" Dec 07 09:22:58 crc kubenswrapper[4838]: E1207 09:22:58.260221 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" podUID="49a96bc4-d983-4088-973e-ebb95fdcc725" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.392436 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" event={"ID":"5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f","Type":"ContainerStarted","Data":"94ee0da516ea7cdf2400a1f8854d7ac66ccd28c59349a2b150415ed29f858723"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.399379 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" event={"ID":"3af86462-0c95-4f05-9e30-c70da589a944","Type":"ContainerStarted","Data":"f154caf1602d99c02ab28783c548327e10674825ec99f4495d088871545cf435"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.413125 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" event={"ID":"7feb13b5-ed7a-47bf-82f1-b749d5b9574e","Type":"ContainerStarted","Data":"f399e787375b3223adc6d57c88ef40bedbb7cd64cd850a41f56a81de2cba5b93"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.413408 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.421492 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" event={"ID":"fdc08e45-9f41-4a0c-8478-10ab2b22ad4e","Type":"ContainerStarted","Data":"085e3123373676b634f66e763f4518686fc88f2301234293e33416376b8587ff"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.421998 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.426099 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.430496 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" event={"ID":"e9af8ce9-a7b0-4bca-b834-2f2f494eb434","Type":"ContainerStarted","Data":"417e37050987a97efd5b1e811bfda7ed36b7d1034b0f8e7366ad8e4c173a3456"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.431041 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.433229 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" event={"ID":"86196566-5b23-4cf1-924e-8a802086e443","Type":"ContainerStarted","Data":"1658d03bf0e86bd6497380ce46cc772d139441fd5255896cd38504e27fb3df5d"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.476454 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" event={"ID":"30216693-f616-448b-b85d-1c6482317ec5","Type":"ContainerStarted","Data":"ed715c6750705bec759a02cf7491e2132e64d662a3bc88e0ab1d106aed2ce7ab"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.477106 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.481217 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" event={"ID":"3c038260-fe42-4c67-8e8c-b994a83c6156","Type":"ContainerStarted","Data":"6cba41540e7501dbeb2d1a9f15d054f9c301c2842f19ee8cebd1145c7b2c45db"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.488425 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" podStartSLOduration=8.26420246 podStartE2EDuration="42.488409171s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.598844052 +0000 UTC m=+955.306163069" lastFinishedPulling="2025-12-07 09:22:52.823050763 +0000 UTC m=+989.530369780" observedRunningTime="2025-12-07 09:22:58.486025203 +0000 UTC m=+995.193344220" watchObservedRunningTime="2025-12-07 09:22:58.488409171 +0000 UTC m=+995.195728188" Dec 07 09:22:58 crc kubenswrapper[4838]: E1207 09:22:58.490692 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" podUID="3c038260-fe42-4c67-8e8c-b994a83c6156" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.494714 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" event={"ID":"5d086115-289a-4478-a657-7ea3927165cd","Type":"ContainerStarted","Data":"24d05444b65f9fa3668de7c5b6109a87d966d2553d63d1bd826aab5aa1222ac7"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.496909 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.520017 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" event={"ID":"ba02a228-91ff-41cf-88ac-b027e697dff9","Type":"ContainerStarted","Data":"e0f319f4c58cd1827688db94ac8d592b9a1c2763ff688ed6910947de3823a4dd"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.539997 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" event={"ID":"49a96bc4-d983-4088-973e-ebb95fdcc725","Type":"ContainerStarted","Data":"f26e9b15e2ba1c8bb9101f66ac934a6b5de1d835fbfee77132b777b2710b553f"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.561088 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" event={"ID":"44803609-79be-4c16-bd10-3d6820d8f24b","Type":"ContainerStarted","Data":"dda9b4b7a5d41849fd73966e350b2ebfa93d6a84cd609aabcbe69235a3134965"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.561710 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.564210 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-hbffh" podStartSLOduration=15.562485567 podStartE2EDuration="42.564187599s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.454353365 +0000 UTC m=+955.161672382" lastFinishedPulling="2025-12-07 09:22:45.456055407 +0000 UTC m=+982.163374414" observedRunningTime="2025-12-07 09:22:58.549571543 +0000 UTC m=+995.256890560" watchObservedRunningTime="2025-12-07 09:22:58.564187599 +0000 UTC m=+995.271506616" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.579097 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" event={"ID":"fac41ae4-04ea-4e07-b672-13fdf35a6530","Type":"ContainerStarted","Data":"9c448ff2ee3d0c494b76f0c903f0da552aa48c6b27e66e72985dda45befb4b20"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.580631 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" event={"ID":"5fc12298-865b-411d-ab04-6cc9d99d8030","Type":"ContainerStarted","Data":"9bfc960f2f2f0a6afe465e535697d4e9afc93fd1fc5124d2eedcef58b9718aad"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.586267 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" event={"ID":"71940e0c-91e5-4634-92d3-d6061fbfc768","Type":"ContainerStarted","Data":"cd4d62259cbfbd3ede38f1d187ddd27377d3809554c55ba8251f1e0c5ae83902"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.586480 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.611220 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" event={"ID":"c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd","Type":"ContainerStarted","Data":"a98b7187590320b76159f27155591c6e4ea7ab4fea2129c18a345e8f9b4a9990"} Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.611260 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.618685 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5cgxx" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.790106 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" podStartSLOduration=15.839628353 podStartE2EDuration="42.790087552s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.363860162 +0000 UTC m=+955.071179189" lastFinishedPulling="2025-12-07 09:22:45.314319371 +0000 UTC m=+982.021638388" observedRunningTime="2025-12-07 09:22:58.631044183 +0000 UTC m=+995.338363190" watchObservedRunningTime="2025-12-07 09:22:58.790087552 +0000 UTC m=+995.497406569" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.791740 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" podStartSLOduration=8.574542467 podStartE2EDuration="42.791730099s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.608148586 +0000 UTC m=+955.315467603" lastFinishedPulling="2025-12-07 09:22:52.825336218 +0000 UTC m=+989.532655235" observedRunningTime="2025-12-07 09:22:58.789075783 +0000 UTC m=+995.496394810" watchObservedRunningTime="2025-12-07 09:22:58.791730099 +0000 UTC m=+995.499049116" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.922244 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" podStartSLOduration=15.453584908 podStartE2EDuration="42.922224695s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:17.846760694 +0000 UTC m=+954.554079711" lastFinishedPulling="2025-12-07 09:22:45.315400481 +0000 UTC m=+982.022719498" observedRunningTime="2025-12-07 09:22:58.877972195 +0000 UTC m=+995.585291212" watchObservedRunningTime="2025-12-07 09:22:58.922224695 +0000 UTC m=+995.629543712" Dec 07 09:22:58 crc kubenswrapper[4838]: I1207 09:22:58.982731 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" podStartSLOduration=8.770036174 podStartE2EDuration="42.982713317s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.6103877 +0000 UTC m=+955.317706717" lastFinishedPulling="2025-12-07 09:22:52.823064843 +0000 UTC m=+989.530383860" observedRunningTime="2025-12-07 09:22:58.978272851 +0000 UTC m=+995.685591868" watchObservedRunningTime="2025-12-07 09:22:58.982713317 +0000 UTC m=+995.690032334" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.046034 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" podStartSLOduration=9.383406771 podStartE2EDuration="43.04601737s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.598695467 +0000 UTC m=+955.306014484" lastFinishedPulling="2025-12-07 09:22:52.261306066 +0000 UTC m=+988.968625083" observedRunningTime="2025-12-07 09:22:59.043680124 +0000 UTC m=+995.750999141" watchObservedRunningTime="2025-12-07 09:22:59.04601737 +0000 UTC m=+995.753336387" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.115114 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" podStartSLOduration=15.426341477 podStartE2EDuration="43.115095107s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:17.766936136 +0000 UTC m=+954.474255153" lastFinishedPulling="2025-12-07 09:22:45.455689766 +0000 UTC m=+982.163008783" observedRunningTime="2025-12-07 09:22:59.106319757 +0000 UTC m=+995.813638774" watchObservedRunningTime="2025-12-07 09:22:59.115095107 +0000 UTC m=+995.822414114" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.668417 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" event={"ID":"fac41ae4-04ea-4e07-b672-13fdf35a6530","Type":"ContainerStarted","Data":"1de99a603aea42c1ef717bff4d8499337bbbad696305725fb43f926e6de60668"} Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.670015 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.683449 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" event={"ID":"5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f","Type":"ContainerStarted","Data":"7fe4c599023c97c44a2abfe0a5457afdccda613b527d5a77bc7ce497afdb5153"} Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.684417 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.707031 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" event={"ID":"f91ec7b3-50b5-4f9f-95a0-c65abeeb879f","Type":"ContainerStarted","Data":"219d07ef9135c0aa7a400425af9dc4c2ef150fa58f68ff65babbc0091332cbd9"} Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.707894 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.708018 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" podStartSLOduration=38.78950079 podStartE2EDuration="43.707997311s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:51.789806679 +0000 UTC m=+988.497125736" lastFinishedPulling="2025-12-07 09:22:56.70830324 +0000 UTC m=+993.415622257" observedRunningTime="2025-12-07 09:22:59.693540369 +0000 UTC m=+996.400859396" watchObservedRunningTime="2025-12-07 09:22:59.707997311 +0000 UTC m=+996.415316328" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.726950 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" event={"ID":"23f05536-d749-43c5-a7cc-7dc63f46e288","Type":"ContainerStarted","Data":"8c2f28c144a3fe3b0ff27bf8a95c541575f3208cd13cb33a6d6dbe975f982b63"} Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.727257 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.731451 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-p5xdf" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.734350 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vb8wb" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.739356 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" podStartSLOduration=3.041539191 podStartE2EDuration="43.739334543s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.537142958 +0000 UTC m=+955.244461975" lastFinishedPulling="2025-12-07 09:22:59.23493831 +0000 UTC m=+995.942257327" observedRunningTime="2025-12-07 09:22:59.728497565 +0000 UTC m=+996.435816582" watchObservedRunningTime="2025-12-07 09:22:59.739334543 +0000 UTC m=+996.446653560" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.749291 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4hn7r" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.752959 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" podStartSLOduration=4.129767583 podStartE2EDuration="43.752945571s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.476747021 +0000 UTC m=+955.184066038" lastFinishedPulling="2025-12-07 09:22:58.099925009 +0000 UTC m=+994.807244026" observedRunningTime="2025-12-07 09:22:59.752618252 +0000 UTC m=+996.459950129" watchObservedRunningTime="2025-12-07 09:22:59.752945571 +0000 UTC m=+996.460264588" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.759975 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-gbbg9" Dec 07 09:22:59 crc kubenswrapper[4838]: I1207 09:22:59.828023 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" podStartSLOduration=4.191239046 podStartE2EDuration="43.827989178s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.44011906 +0000 UTC m=+955.147438077" lastFinishedPulling="2025-12-07 09:22:58.076869192 +0000 UTC m=+994.784188209" observedRunningTime="2025-12-07 09:22:59.792305402 +0000 UTC m=+996.499624419" watchObservedRunningTime="2025-12-07 09:22:59.827989178 +0000 UTC m=+996.535308195" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.752535 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" event={"ID":"86196566-5b23-4cf1-924e-8a802086e443","Type":"ContainerStarted","Data":"81648833e0f00b44a165ef7edd61f69fb5414d65d31ed997e3543a18b6a17021"} Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.752875 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.759899 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" event={"ID":"5fc12298-865b-411d-ab04-6cc9d99d8030","Type":"ContainerStarted","Data":"dff756fb664c768ecc2482819ceded5017eed1ba81ae471024598ab2ae615a1f"} Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.760650 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.768862 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" event={"ID":"3af86462-0c95-4f05-9e30-c70da589a944","Type":"ContainerStarted","Data":"ba50dfd6031876fdca6eec3d6a790ea1c2362a0ed9bde656d86d9fadde463c3d"} Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.769519 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.771323 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" event={"ID":"ba02a228-91ff-41cf-88ac-b027e697dff9","Type":"ContainerStarted","Data":"0b42f55b00ab12d9c31fc594f7213901bf98d708cf9c9f3140ecb451a1535d7a"} Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.771746 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.777749 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" event={"ID":"49a96bc4-d983-4088-973e-ebb95fdcc725","Type":"ContainerStarted","Data":"3b444eb8b674163ae40a00c8d9da227cfadb55ffd150b526ace5cbc25b98d599"} Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.777785 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.786040 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" podStartSLOduration=3.850263929 podStartE2EDuration="44.786024919s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.565411661 +0000 UTC m=+955.272730678" lastFinishedPulling="2025-12-07 09:22:59.501172651 +0000 UTC m=+996.208491668" observedRunningTime="2025-12-07 09:23:00.785997518 +0000 UTC m=+997.493316535" watchObservedRunningTime="2025-12-07 09:23:00.786024919 +0000 UTC m=+997.493343936" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.837374 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" podStartSLOduration=4.137822959 podStartE2EDuration="44.83735944s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.536747817 +0000 UTC m=+955.244066834" lastFinishedPulling="2025-12-07 09:22:59.236284298 +0000 UTC m=+995.943603315" observedRunningTime="2025-12-07 09:23:00.833705576 +0000 UTC m=+997.541024593" watchObservedRunningTime="2025-12-07 09:23:00.83735944 +0000 UTC m=+997.544678457" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.884338 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" podStartSLOduration=4.209278532 podStartE2EDuration="44.884317008s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.564954388 +0000 UTC m=+955.272273405" lastFinishedPulling="2025-12-07 09:22:59.239992864 +0000 UTC m=+995.947311881" observedRunningTime="2025-12-07 09:23:00.879181381 +0000 UTC m=+997.586500398" watchObservedRunningTime="2025-12-07 09:23:00.884317008 +0000 UTC m=+997.591636025" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.884678 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" podStartSLOduration=3.353723641 podStartE2EDuration="44.884671018s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:17.968496185 +0000 UTC m=+954.675815192" lastFinishedPulling="2025-12-07 09:22:59.499443552 +0000 UTC m=+996.206762569" observedRunningTime="2025-12-07 09:23:00.857805403 +0000 UTC m=+997.565124420" watchObservedRunningTime="2025-12-07 09:23:00.884671018 +0000 UTC m=+997.591990035" Dec 07 09:23:00 crc kubenswrapper[4838]: I1207 09:23:00.931502 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" podStartSLOduration=4.159317455 podStartE2EDuration="44.931479841s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.466430228 +0000 UTC m=+955.173749245" lastFinishedPulling="2025-12-07 09:22:59.238592614 +0000 UTC m=+995.945911631" observedRunningTime="2025-12-07 09:23:00.898271815 +0000 UTC m=+997.605590822" watchObservedRunningTime="2025-12-07 09:23:00.931479841 +0000 UTC m=+997.638798858" Dec 07 09:23:02 crc kubenswrapper[4838]: E1207 09:23:02.621956 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" podUID="700b9dd5-3c10-4fa7-bde4-0c9a4f778608" Dec 07 09:23:05 crc kubenswrapper[4838]: I1207 09:23:05.810312 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" event={"ID":"07975bb2-d979-4265-bba2-1254d58b7267","Type":"ContainerStarted","Data":"4d2c409681c000c3bb5a93df521cb4f7ac540232871945e804048b81d341a005"} Dec 07 09:23:05 crc kubenswrapper[4838]: I1207 09:23:05.810809 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" event={"ID":"07975bb2-d979-4265-bba2-1254d58b7267","Type":"ContainerStarted","Data":"b4c321d3a1fa4d5b73f83a843e06bbb09160d09a335c62e594c785675cd4fc5e"} Dec 07 09:23:05 crc kubenswrapper[4838]: I1207 09:23:05.811536 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:23:05 crc kubenswrapper[4838]: I1207 09:23:05.813458 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" event={"ID":"3bd0361d-590b-4a06-b32c-6f7ceb427fbd","Type":"ContainerStarted","Data":"9e533a8735b37efbba2281ba013c5b555614555f68f88fec2d3521649ea5890a"} Dec 07 09:23:05 crc kubenswrapper[4838]: I1207 09:23:05.813482 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" event={"ID":"3bd0361d-590b-4a06-b32c-6f7ceb427fbd","Type":"ContainerStarted","Data":"f68a365e668acd948623f36b65d19d6863700f90d1e61b57e3437666f6d4d197"} Dec 07 09:23:05 crc kubenswrapper[4838]: I1207 09:23:05.813917 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" Dec 07 09:23:05 crc kubenswrapper[4838]: I1207 09:23:05.843083 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" podStartSLOduration=39.383855005 podStartE2EDuration="49.843067415s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:54.639427286 +0000 UTC m=+991.346746303" lastFinishedPulling="2025-12-07 09:23:05.098639696 +0000 UTC m=+1001.805958713" observedRunningTime="2025-12-07 09:23:05.837476896 +0000 UTC m=+1002.544795913" watchObservedRunningTime="2025-12-07 09:23:05.843067415 +0000 UTC m=+1002.550386432" Dec 07 09:23:06 crc kubenswrapper[4838]: I1207 09:23:06.670199 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2ftqb" Dec 07 09:23:06 crc kubenswrapper[4838]: I1207 09:23:06.689700 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" podStartSLOduration=4.154228605 podStartE2EDuration="50.689667023s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.565790072 +0000 UTC m=+955.273109089" lastFinishedPulling="2025-12-07 09:23:05.10122849 +0000 UTC m=+1001.808547507" observedRunningTime="2025-12-07 09:23:05.87273305 +0000 UTC m=+1002.580052057" watchObservedRunningTime="2025-12-07 09:23:06.689667023 +0000 UTC m=+1003.396986040" Dec 07 09:23:06 crc kubenswrapper[4838]: I1207 09:23:06.797923 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-pmdl9" Dec 07 09:23:06 crc kubenswrapper[4838]: I1207 09:23:06.848614 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-9hg7c" Dec 07 09:23:06 crc kubenswrapper[4838]: I1207 09:23:06.849710 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-gqrds" Dec 07 09:23:07 crc kubenswrapper[4838]: I1207 09:23:07.008403 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9cwrh" Dec 07 09:23:07 crc kubenswrapper[4838]: I1207 09:23:07.047991 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-t6tz7" Dec 07 09:23:07 crc kubenswrapper[4838]: I1207 09:23:07.188318 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-6vrx6" Dec 07 09:23:07 crc kubenswrapper[4838]: I1207 09:23:07.225294 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-b29lr" Dec 07 09:23:07 crc kubenswrapper[4838]: I1207 09:23:07.290039 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9jtsq" Dec 07 09:23:07 crc kubenswrapper[4838]: I1207 09:23:07.391174 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-54jxx" Dec 07 09:23:07 crc kubenswrapper[4838]: I1207 09:23:07.625183 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wz9hz" Dec 07 09:23:09 crc kubenswrapper[4838]: I1207 09:23:09.362675 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5b47684954-7mp9q" Dec 07 09:23:11 crc kubenswrapper[4838]: I1207 09:23:11.852615 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" event={"ID":"3c038260-fe42-4c67-8e8c-b994a83c6156","Type":"ContainerStarted","Data":"8bab37393344a12fac2d796075963c6bf4953eaeb9eb50b15c7e5bb916fd6d16"} Dec 07 09:23:11 crc kubenswrapper[4838]: I1207 09:23:11.853122 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" Dec 07 09:23:11 crc kubenswrapper[4838]: I1207 09:23:11.868987 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" podStartSLOduration=3.284280762 podStartE2EDuration="55.86896298s" podCreationTimestamp="2025-12-07 09:22:16 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.570843386 +0000 UTC m=+955.278162403" lastFinishedPulling="2025-12-07 09:23:11.155525594 +0000 UTC m=+1007.862844621" observedRunningTime="2025-12-07 09:23:11.865376368 +0000 UTC m=+1008.572695385" watchObservedRunningTime="2025-12-07 09:23:11.86896298 +0000 UTC m=+1008.576281997" Dec 07 09:23:12 crc kubenswrapper[4838]: I1207 09:23:12.632888 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4q7pd" Dec 07 09:23:16 crc kubenswrapper[4838]: I1207 09:23:16.898552 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" event={"ID":"700b9dd5-3c10-4fa7-bde4-0c9a4f778608","Type":"ContainerStarted","Data":"8400d5954b7e7e21a150386071ab625cb723495baa350897df12185aec9131d0"} Dec 07 09:23:16 crc kubenswrapper[4838]: I1207 09:23:16.920198 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tl2fn" podStartSLOduration=2.504287316 podStartE2EDuration="59.920166621s" podCreationTimestamp="2025-12-07 09:22:17 +0000 UTC" firstStartedPulling="2025-12-07 09:22:18.599155541 +0000 UTC m=+955.306474558" lastFinishedPulling="2025-12-07 09:23:16.015034846 +0000 UTC m=+1012.722353863" observedRunningTime="2025-12-07 09:23:16.91347008 +0000 UTC m=+1013.620789117" watchObservedRunningTime="2025-12-07 09:23:16.920166621 +0000 UTC m=+1013.627485638" Dec 07 09:23:16 crc kubenswrapper[4838]: I1207 09:23:16.940726 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6cwcf" Dec 07 09:23:17 crc kubenswrapper[4838]: I1207 09:23:17.646670 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-v5hzl" Dec 07 09:23:19 crc kubenswrapper[4838]: I1207 09:23:19.048124 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-744f8cb766qxndm" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.156531 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-68szs"] Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.158306 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.169683 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.169809 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-xdpg9" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.169953 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.170185 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.193548 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-68szs"] Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.195797 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-config\") pod \"dnsmasq-dns-5cd484bb89-68szs\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.195907 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb5s5\" (UniqueName: \"kubernetes.io/projected/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-kube-api-access-mb5s5\") pod \"dnsmasq-dns-5cd484bb89-68szs\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.228876 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567c455747-wnfnj"] Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.230308 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.236093 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.256527 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-wnfnj"] Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.296760 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6xkd\" (UniqueName: \"kubernetes.io/projected/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-kube-api-access-n6xkd\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.296827 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-config\") pod \"dnsmasq-dns-5cd484bb89-68szs\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.296857 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-config\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.296891 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-dns-svc\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.296919 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb5s5\" (UniqueName: \"kubernetes.io/projected/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-kube-api-access-mb5s5\") pod \"dnsmasq-dns-5cd484bb89-68szs\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.297949 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-config\") pod \"dnsmasq-dns-5cd484bb89-68szs\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.346696 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb5s5\" (UniqueName: \"kubernetes.io/projected/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-kube-api-access-mb5s5\") pod \"dnsmasq-dns-5cd484bb89-68szs\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.397743 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6xkd\" (UniqueName: \"kubernetes.io/projected/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-kube-api-access-n6xkd\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.397796 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-config\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.397842 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-dns-svc\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.398643 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-dns-svc\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.399387 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-config\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.426489 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6xkd\" (UniqueName: \"kubernetes.io/projected/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-kube-api-access-n6xkd\") pod \"dnsmasq-dns-567c455747-wnfnj\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.478985 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.555067 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:23:35 crc kubenswrapper[4838]: I1207 09:23:35.977001 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-68szs"] Dec 07 09:23:35 crc kubenswrapper[4838]: W1207 09:23:35.980199 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0256fa90_d1c3_4ef9_ab26_2e6fde11b77a.slice/crio-3c25f3ec119bfc2a5727f8817a2f8e6ad13ff8ef8f99109610a2814a0b983f4d WatchSource:0}: Error finding container 3c25f3ec119bfc2a5727f8817a2f8e6ad13ff8ef8f99109610a2814a0b983f4d: Status 404 returned error can't find the container with id 3c25f3ec119bfc2a5727f8817a2f8e6ad13ff8ef8f99109610a2814a0b983f4d Dec 07 09:23:36 crc kubenswrapper[4838]: I1207 09:23:36.027708 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-68szs" event={"ID":"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a","Type":"ContainerStarted","Data":"3c25f3ec119bfc2a5727f8817a2f8e6ad13ff8ef8f99109610a2814a0b983f4d"} Dec 07 09:23:36 crc kubenswrapper[4838]: I1207 09:23:36.063183 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-wnfnj"] Dec 07 09:23:36 crc kubenswrapper[4838]: W1207 09:23:36.069634 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd34e8e3d_9e9c_4029_b0e2_a5ad7a5990b0.slice/crio-573e1b7bf90574564572abf30ca9cdc92b527f96216e36123df978a917b7bc31 WatchSource:0}: Error finding container 573e1b7bf90574564572abf30ca9cdc92b527f96216e36123df978a917b7bc31: Status 404 returned error can't find the container with id 573e1b7bf90574564572abf30ca9cdc92b527f96216e36123df978a917b7bc31 Dec 07 09:23:37 crc kubenswrapper[4838]: I1207 09:23:37.040953 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-wnfnj" event={"ID":"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0","Type":"ContainerStarted","Data":"573e1b7bf90574564572abf30ca9cdc92b527f96216e36123df978a917b7bc31"} Dec 07 09:23:37 crc kubenswrapper[4838]: I1207 09:23:37.883124 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-wnfnj"] Dec 07 09:23:37 crc kubenswrapper[4838]: I1207 09:23:37.917310 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-859d485f47-b6n8d"] Dec 07 09:23:37 crc kubenswrapper[4838]: I1207 09:23:37.918697 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:37 crc kubenswrapper[4838]: I1207 09:23:37.938478 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-b6n8d"] Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.040290 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-dns-svc\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.040361 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-config\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.040455 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlvv9\" (UniqueName: \"kubernetes.io/projected/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-kube-api-access-tlvv9\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.142037 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-dns-svc\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.142080 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-config\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.142140 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlvv9\" (UniqueName: \"kubernetes.io/projected/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-kube-api-access-tlvv9\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.142948 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-config\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.143485 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-dns-svc\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.181165 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlvv9\" (UniqueName: \"kubernetes.io/projected/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-kube-api-access-tlvv9\") pod \"dnsmasq-dns-859d485f47-b6n8d\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.256290 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.348433 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-68szs"] Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.402607 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb666b895-sgg8c"] Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.403686 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.435189 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-sgg8c"] Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.552542 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-config\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.552619 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7pdp\" (UniqueName: \"kubernetes.io/projected/280b54c8-5174-4b27-b41e-b7bd28b27111-kube-api-access-j7pdp\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.552666 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-dns-svc\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.653514 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-config\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.653874 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7pdp\" (UniqueName: \"kubernetes.io/projected/280b54c8-5174-4b27-b41e-b7bd28b27111-kube-api-access-j7pdp\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.653914 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-dns-svc\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.654797 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-dns-svc\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.655359 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-config\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.704874 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7pdp\" (UniqueName: \"kubernetes.io/projected/280b54c8-5174-4b27-b41e-b7bd28b27111-kube-api-access-j7pdp\") pod \"dnsmasq-dns-cb666b895-sgg8c\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:38 crc kubenswrapper[4838]: I1207 09:23:38.732127 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.025783 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-b6n8d"] Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.084024 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.085218 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.089390 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.089597 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.094354 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.094539 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.094654 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-875vw" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.094773 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.099086 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.101642 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:23:39 crc kubenswrapper[4838]: W1207 09:23:39.214318 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d37f7b3_d058_4f8f_8fb4_f7e6e0da552e.slice/crio-1603d085eb920214d08b0edb5a43658cbad30737fc98e958a96279e673453e7e WatchSource:0}: Error finding container 1603d085eb920214d08b0edb5a43658cbad30737fc98e958a96279e673453e7e: Status 404 returned error can't find the container with id 1603d085eb920214d08b0edb5a43658cbad30737fc98e958a96279e673453e7e Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270311 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270389 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270420 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270464 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270482 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzmbr\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-kube-api-access-qzmbr\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270549 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270570 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270645 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270681 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270705 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.270762 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373095 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373141 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373164 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373181 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373207 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373247 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373282 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373297 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373318 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373339 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.373361 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzmbr\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-kube-api-access-qzmbr\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.374695 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.375347 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.375453 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.375900 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.376170 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.376517 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.399681 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.400155 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.404208 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzmbr\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-kube-api-access-qzmbr\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.405526 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.407582 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.409553 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.421479 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.511984 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-sgg8c"] Dec 07 09:23:39 crc kubenswrapper[4838]: W1207 09:23:39.525268 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod280b54c8_5174_4b27_b41e_b7bd28b27111.slice/crio-4f9420f4efe2bc8cd64d47373180ac374e82e9fd3feb11d90eb706e162bd9b9c WatchSource:0}: Error finding container 4f9420f4efe2bc8cd64d47373180ac374e82e9fd3feb11d90eb706e162bd9b9c: Status 404 returned error can't find the container with id 4f9420f4efe2bc8cd64d47373180ac374e82e9fd3feb11d90eb706e162bd9b9c Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.538348 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.543973 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.551002 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.551192 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.551337 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.551509 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nxdd4" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.551517 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.551634 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.552076 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.562062 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.685677 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eae80b29-4c64-403b-be38-7fd30b7ba66f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.685768 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.685791 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.685861 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4pn6\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-kube-api-access-m4pn6\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.685902 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.685933 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.685996 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.686033 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eae80b29-4c64-403b-be38-7fd30b7ba66f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.686076 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.686110 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-config-data\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.686197 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787570 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787633 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eae80b29-4c64-403b-be38-7fd30b7ba66f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787662 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787678 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787701 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4pn6\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-kube-api-access-m4pn6\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787718 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787740 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787767 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787787 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eae80b29-4c64-403b-be38-7fd30b7ba66f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787807 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.787838 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-config-data\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.788684 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-config-data\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.788996 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.789668 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.793368 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.793591 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.793655 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.800462 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.820438 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eae80b29-4c64-403b-be38-7fd30b7ba66f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.821021 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eae80b29-4c64-403b-be38-7fd30b7ba66f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.821503 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.837712 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4pn6\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-kube-api-access-m4pn6\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.846998 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " pod="openstack/rabbitmq-server-0" Dec 07 09:23:39 crc kubenswrapper[4838]: I1207 09:23:39.882088 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.016597 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:23:40 crc kubenswrapper[4838]: W1207 09:23:40.023809 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20d0a08e_11bd_4cb9_b5c7_b61abbbd4454.slice/crio-1684f067ea189296b647a867168cc7d63950375ad7c8de7aaeac4327c4a6d034 WatchSource:0}: Error finding container 1684f067ea189296b647a867168cc7d63950375ad7c8de7aaeac4327c4a6d034: Status 404 returned error can't find the container with id 1684f067ea189296b647a867168cc7d63950375ad7c8de7aaeac4327c4a6d034 Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.095614 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" event={"ID":"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e","Type":"ContainerStarted","Data":"1603d085eb920214d08b0edb5a43658cbad30737fc98e958a96279e673453e7e"} Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.110394 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" event={"ID":"280b54c8-5174-4b27-b41e-b7bd28b27111","Type":"ContainerStarted","Data":"4f9420f4efe2bc8cd64d47373180ac374e82e9fd3feb11d90eb706e162bd9b9c"} Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.115248 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454","Type":"ContainerStarted","Data":"1684f067ea189296b647a867168cc7d63950375ad7c8de7aaeac4327c4a6d034"} Dec 07 09:23:40 crc kubenswrapper[4838]: W1207 09:23:40.399077 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeae80b29_4c64_403b_be38_7fd30b7ba66f.slice/crio-19de77fd1c4624bfd4eab30715225b42a8401b4eb1bbdeaf0aa3b4cc814e2316 WatchSource:0}: Error finding container 19de77fd1c4624bfd4eab30715225b42a8401b4eb1bbdeaf0aa3b4cc814e2316: Status 404 returned error can't find the container with id 19de77fd1c4624bfd4eab30715225b42a8401b4eb1bbdeaf0aa3b4cc814e2316 Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.403147 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.775221 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.779776 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.782146 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gcldr" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.782559 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.782680 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.783276 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.784191 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.787377 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.943806 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-kolla-config\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.944234 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.944342 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.944395 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d111910-1e1a-4b37-87c2-5c8f55110a73-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.944425 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d111910-1e1a-4b37-87c2-5c8f55110a73-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.944526 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-config-data-default\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.944644 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv687\" (UniqueName: \"kubernetes.io/projected/3d111910-1e1a-4b37-87c2-5c8f55110a73-kube-api-access-tv687\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:40 crc kubenswrapper[4838]: I1207 09:23:40.944698 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d111910-1e1a-4b37-87c2-5c8f55110a73-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.045761 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d111910-1e1a-4b37-87c2-5c8f55110a73-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046205 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d111910-1e1a-4b37-87c2-5c8f55110a73-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046227 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-config-data-default\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046280 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv687\" (UniqueName: \"kubernetes.io/projected/3d111910-1e1a-4b37-87c2-5c8f55110a73-kube-api-access-tv687\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046316 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d111910-1e1a-4b37-87c2-5c8f55110a73-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046351 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-kolla-config\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046380 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046413 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046750 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.046845 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d111910-1e1a-4b37-87c2-5c8f55110a73-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.047871 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-config-data-default\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.048257 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-kolla-config\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.049307 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d111910-1e1a-4b37-87c2-5c8f55110a73-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.072391 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d111910-1e1a-4b37-87c2-5c8f55110a73-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.078307 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d111910-1e1a-4b37-87c2-5c8f55110a73-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.086388 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv687\" (UniqueName: \"kubernetes.io/projected/3d111910-1e1a-4b37-87c2-5c8f55110a73-kube-api-access-tv687\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.097802 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3d111910-1e1a-4b37-87c2-5c8f55110a73\") " pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.131133 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 07 09:23:41 crc kubenswrapper[4838]: I1207 09:23:41.149776 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eae80b29-4c64-403b-be38-7fd30b7ba66f","Type":"ContainerStarted","Data":"19de77fd1c4624bfd4eab30715225b42a8401b4eb1bbdeaf0aa3b4cc814e2316"} Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.035669 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.105283 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.107972 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.113032 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.113404 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.113508 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jcknn" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.113597 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.119823 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.172341 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d111910-1e1a-4b37-87c2-5c8f55110a73","Type":"ContainerStarted","Data":"e97c19eaf70da930a27083cc38ba1ea114c52f0910a476369da94e0fbf99d78c"} Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.287051 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.287182 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.288617 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.288685 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.288767 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rs8j\" (UniqueName: \"kubernetes.io/projected/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-kube-api-access-5rs8j\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.288809 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.288847 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.288875 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390258 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rs8j\" (UniqueName: \"kubernetes.io/projected/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-kube-api-access-5rs8j\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390310 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390329 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390366 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390397 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390448 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390475 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.390500 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.391127 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.391711 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.392772 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.394492 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.394671 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.418508 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.424191 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rs8j\" (UniqueName: \"kubernetes.io/projected/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-kube-api-access-5rs8j\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.433000 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.445368 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c\") " pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.455924 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.531662 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.532556 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.537212 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.537404 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.537532 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-rh9ts" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.634001 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.701081 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76cc3a82-a620-435a-99f0-9a91297d2ee2-config-data\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.701440 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85n2p\" (UniqueName: \"kubernetes.io/projected/76cc3a82-a620-435a-99f0-9a91297d2ee2-kube-api-access-85n2p\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.701509 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76cc3a82-a620-435a-99f0-9a91297d2ee2-kolla-config\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.701526 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/76cc3a82-a620-435a-99f0-9a91297d2ee2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.701571 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76cc3a82-a620-435a-99f0-9a91297d2ee2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.802742 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76cc3a82-a620-435a-99f0-9a91297d2ee2-config-data\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.802875 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85n2p\" (UniqueName: \"kubernetes.io/projected/76cc3a82-a620-435a-99f0-9a91297d2ee2-kube-api-access-85n2p\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.802939 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76cc3a82-a620-435a-99f0-9a91297d2ee2-kolla-config\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.802961 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/76cc3a82-a620-435a-99f0-9a91297d2ee2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.802988 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76cc3a82-a620-435a-99f0-9a91297d2ee2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.806059 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/76cc3a82-a620-435a-99f0-9a91297d2ee2-kolla-config\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.822217 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76cc3a82-a620-435a-99f0-9a91297d2ee2-config-data\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.822656 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/76cc3a82-a620-435a-99f0-9a91297d2ee2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.823513 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76cc3a82-a620-435a-99f0-9a91297d2ee2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.832827 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85n2p\" (UniqueName: \"kubernetes.io/projected/76cc3a82-a620-435a-99f0-9a91297d2ee2-kube-api-access-85n2p\") pod \"memcached-0\" (UID: \"76cc3a82-a620-435a-99f0-9a91297d2ee2\") " pod="openstack/memcached-0" Dec 07 09:23:42 crc kubenswrapper[4838]: I1207 09:23:42.907331 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 07 09:23:43 crc kubenswrapper[4838]: I1207 09:23:43.128567 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 07 09:23:43 crc kubenswrapper[4838]: I1207 09:23:43.203063 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c","Type":"ContainerStarted","Data":"88c1a70a561656db3c7ad6f2baeea100caa8be8e59a5951bd7632951313210e3"} Dec 07 09:23:43 crc kubenswrapper[4838]: I1207 09:23:43.584234 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.504754 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.505738 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.509649 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.537213 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-46tgh" Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.642039 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwn9b\" (UniqueName: \"kubernetes.io/projected/be998839-55b0-48de-9903-197e3684a4cf-kube-api-access-mwn9b\") pod \"kube-state-metrics-0\" (UID: \"be998839-55b0-48de-9903-197e3684a4cf\") " pod="openstack/kube-state-metrics-0" Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.743387 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwn9b\" (UniqueName: \"kubernetes.io/projected/be998839-55b0-48de-9903-197e3684a4cf-kube-api-access-mwn9b\") pod \"kube-state-metrics-0\" (UID: \"be998839-55b0-48de-9903-197e3684a4cf\") " pod="openstack/kube-state-metrics-0" Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.782900 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwn9b\" (UniqueName: \"kubernetes.io/projected/be998839-55b0-48de-9903-197e3684a4cf-kube-api-access-mwn9b\") pod \"kube-state-metrics-0\" (UID: \"be998839-55b0-48de-9903-197e3684a4cf\") " pod="openstack/kube-state-metrics-0" Dec 07 09:23:44 crc kubenswrapper[4838]: I1207 09:23:44.870947 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.088767 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sbl7t"] Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.089928 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.094850 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.095561 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.095759 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-t67mp" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.101534 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbl7t"] Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.128706 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-run\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.128775 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-run-ovn\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.128803 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ae58f2-6cd2-4782-96b1-a88597b25b37-combined-ca-bundle\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.128993 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-log-ovn\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.129123 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3ae58f2-6cd2-4782-96b1-a88597b25b37-scripts\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.129198 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hffh\" (UniqueName: \"kubernetes.io/projected/a3ae58f2-6cd2-4782-96b1-a88597b25b37-kube-api-access-4hffh\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.129241 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ae58f2-6cd2-4782-96b1-a88597b25b37-ovn-controller-tls-certs\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.159282 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-gb65g"] Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.161373 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.193088 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gb65g"] Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.233064 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-run-ovn\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.233113 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ae58f2-6cd2-4782-96b1-a88597b25b37-combined-ca-bundle\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.233478 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-log-ovn\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.233550 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3ae58f2-6cd2-4782-96b1-a88597b25b37-scripts\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.233657 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hffh\" (UniqueName: \"kubernetes.io/projected/a3ae58f2-6cd2-4782-96b1-a88597b25b37-kube-api-access-4hffh\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.233674 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ae58f2-6cd2-4782-96b1-a88597b25b37-ovn-controller-tls-certs\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.233697 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-run\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.243394 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3ae58f2-6cd2-4782-96b1-a88597b25b37-scripts\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.243606 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-run-ovn\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.243676 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-log-ovn\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.244624 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3ae58f2-6cd2-4782-96b1-a88597b25b37-combined-ca-bundle\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.245988 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a3ae58f2-6cd2-4782-96b1-a88597b25b37-var-run\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.256503 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hffh\" (UniqueName: \"kubernetes.io/projected/a3ae58f2-6cd2-4782-96b1-a88597b25b37-kube-api-access-4hffh\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.281391 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3ae58f2-6cd2-4782-96b1-a88597b25b37-ovn-controller-tls-certs\") pod \"ovn-controller-sbl7t\" (UID: \"a3ae58f2-6cd2-4782-96b1-a88597b25b37\") " pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.339424 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-log\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.339489 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-etc-ovs\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.339536 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc745e13-8d6d-4e12-a733-5c081c0e4283-scripts\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.339639 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-lib\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.339681 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shxff\" (UniqueName: \"kubernetes.io/projected/dc745e13-8d6d-4e12-a733-5c081c0e4283-kube-api-access-shxff\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.339745 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-run\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443172 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-lib\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443237 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shxff\" (UniqueName: \"kubernetes.io/projected/dc745e13-8d6d-4e12-a733-5c081c0e4283-kube-api-access-shxff\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443289 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-run\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443317 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-log\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443334 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-etc-ovs\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443399 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc745e13-8d6d-4e12-a733-5c081c0e4283-scripts\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443570 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-lib\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443719 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-etc-ovs\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443715 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-run\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.443731 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/dc745e13-8d6d-4e12-a733-5c081c0e4283-var-log\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.445585 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc745e13-8d6d-4e12-a733-5c081c0e4283-scripts\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.466716 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.471373 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shxff\" (UniqueName: \"kubernetes.io/projected/dc745e13-8d6d-4e12-a733-5c081c0e4283-kube-api-access-shxff\") pod \"ovn-controller-ovs-gb65g\" (UID: \"dc745e13-8d6d-4e12-a733-5c081c0e4283\") " pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:48 crc kubenswrapper[4838]: I1207 09:23:48.499327 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.723643 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.725045 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.727027 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wrvfm" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.727618 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.728060 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.728255 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.732138 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.848196 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.888688 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.888748 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvddl\" (UniqueName: \"kubernetes.io/projected/cec74e7b-a2aa-46a0-b318-61decd213728-kube-api-access-zvddl\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.888870 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.888905 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cec74e7b-a2aa-46a0-b318-61decd213728-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.888935 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.888967 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec74e7b-a2aa-46a0-b318-61decd213728-config\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.889054 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec74e7b-a2aa-46a0-b318-61decd213728-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.889084 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.990751 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.990806 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cec74e7b-a2aa-46a0-b318-61decd213728-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.990840 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.990857 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec74e7b-a2aa-46a0-b318-61decd213728-config\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.990907 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec74e7b-a2aa-46a0-b318-61decd213728-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.990925 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.990986 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.991002 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvddl\" (UniqueName: \"kubernetes.io/projected/cec74e7b-a2aa-46a0-b318-61decd213728-kube-api-access-zvddl\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.991134 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.991902 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cec74e7b-a2aa-46a0-b318-61decd213728-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.992735 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec74e7b-a2aa-46a0-b318-61decd213728-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:50 crc kubenswrapper[4838]: I1207 09:23:50.995606 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec74e7b-a2aa-46a0-b318-61decd213728-config\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.001387 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.010480 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.011804 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvddl\" (UniqueName: \"kubernetes.io/projected/cec74e7b-a2aa-46a0-b318-61decd213728-kube-api-access-zvddl\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.012233 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cec74e7b-a2aa-46a0-b318-61decd213728-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.030075 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cec74e7b-a2aa-46a0-b318-61decd213728\") " pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.041352 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.864287 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.865595 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.868777 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.869021 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.872141 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.873097 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-4892h" Dec 07 09:23:51 crc kubenswrapper[4838]: I1207 09:23:51.881084 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.006925 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.006975 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.006996 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.007024 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.007059 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.007072 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-config\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.007089 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrkwg\" (UniqueName: \"kubernetes.io/projected/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-kube-api-access-vrkwg\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.007107 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.108868 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.108929 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.108949 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.108996 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.109031 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.109046 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-config\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.109080 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrkwg\" (UniqueName: \"kubernetes.io/projected/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-kube-api-access-vrkwg\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.109098 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.109304 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.110334 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-config\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.110385 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.110453 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.112779 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.114872 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.118419 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.136583 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.146149 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrkwg\" (UniqueName: \"kubernetes.io/projected/e06fb79a-09b7-42af-bb04-1ef9ba3f7560-kube-api-access-vrkwg\") pod \"ovsdbserver-sb-0\" (UID: \"e06fb79a-09b7-42af-bb04-1ef9ba3f7560\") " pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.225650 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 07 09:23:52 crc kubenswrapper[4838]: I1207 09:23:52.293069 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"76cc3a82-a620-435a-99f0-9a91297d2ee2","Type":"ContainerStarted","Data":"34e6c616a0250f351abb57947c2627c2412bc0d4951b61641157f0cbb5628562"} Dec 07 09:24:00 crc kubenswrapper[4838]: E1207 09:24:00.067178 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 07 09:24:00 crc kubenswrapper[4838]: E1207 09:24:00.068161 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m4pn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(eae80b29-4c64-403b-be38-7fd30b7ba66f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:00 crc kubenswrapper[4838]: E1207 09:24:00.069403 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" Dec 07 09:24:00 crc kubenswrapper[4838]: E1207 09:24:00.351783 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-server-0" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" Dec 07 09:24:08 crc kubenswrapper[4838]: E1207 09:24:08.497289 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 07 09:24:08 crc kubenswrapper[4838]: E1207 09:24:08.498040 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qzmbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(20d0a08e-11bd-4cb9-b5c7-b61abbbd4454): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:08 crc kubenswrapper[4838]: E1207 09:24:08.499458 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" Dec 07 09:24:08 crc kubenswrapper[4838]: I1207 09:24:08.962419 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbl7t"] Dec 07 09:24:09 crc kubenswrapper[4838]: I1207 09:24:09.181775 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 07 09:24:09 crc kubenswrapper[4838]: E1207 09:24:09.419889 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" Dec 07 09:24:11 crc kubenswrapper[4838]: E1207 09:24:11.406926 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 07 09:24:11 crc kubenswrapper[4838]: E1207 09:24:11.408176 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tv687,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(3d111910-1e1a-4b37-87c2-5c8f55110a73): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:11 crc kubenswrapper[4838]: E1207 09:24:11.409436 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="3d111910-1e1a-4b37-87c2-5c8f55110a73" Dec 07 09:24:11 crc kubenswrapper[4838]: E1207 09:24:11.434623 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-galera-0" podUID="3d111910-1e1a-4b37-87c2-5c8f55110a73" Dec 07 09:24:11 crc kubenswrapper[4838]: E1207 09:24:11.437520 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 07 09:24:11 crc kubenswrapper[4838]: E1207 09:24:11.437662 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5rs8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:11 crc kubenswrapper[4838]: E1207 09:24:11.442997 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.199036 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.199626 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n6xkd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-567c455747-wnfnj_openstack(d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.201082 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-567c455747-wnfnj" podUID="d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.207372 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.207636 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j7pdp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-cb666b895-sgg8c_openstack(280b54c8-5174-4b27-b41e-b7bd28b27111): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.208964 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.446372 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" Dec 07 09:24:12 crc kubenswrapper[4838]: E1207 09:24:12.446847 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c" Dec 07 09:24:13 crc kubenswrapper[4838]: E1207 09:24:13.236778 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 07 09:24:13 crc kubenswrapper[4838]: E1207 09:24:13.237166 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tlvv9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-859d485f47-b6n8d_openstack(0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:13 crc kubenswrapper[4838]: E1207 09:24:13.237743 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 07 09:24:13 crc kubenswrapper[4838]: E1207 09:24:13.238032 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mb5s5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5cd484bb89-68szs_openstack(0256fa90-d1c3-4ef9-ab26-2e6fde11b77a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:24:13 crc kubenswrapper[4838]: E1207 09:24:13.239220 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" Dec 07 09:24:13 crc kubenswrapper[4838]: E1207 09:24:13.239273 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5cd484bb89-68szs" podUID="0256fa90-d1c3-4ef9-ab26-2e6fde11b77a" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.351506 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.454657 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cec74e7b-a2aa-46a0-b318-61decd213728","Type":"ContainerStarted","Data":"699900913e71cfd26788b5d4c7214242533b5ef7696e2997b382d75052a1ac09"} Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.456432 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-wnfnj" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.456456 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-wnfnj" event={"ID":"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0","Type":"ContainerDied","Data":"573e1b7bf90574564572abf30ca9cdc92b527f96216e36123df978a917b7bc31"} Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.457604 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t" event={"ID":"a3ae58f2-6cd2-4782-96b1-a88597b25b37","Type":"ContainerStarted","Data":"4dcb8d286e77b2a3709d029df5ee15ccb3718899c2957fac60fa0924d0e7112a"} Dec 07 09:24:13 crc kubenswrapper[4838]: E1207 09:24:13.459547 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.498516 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-dns-svc\") pod \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.498746 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6xkd\" (UniqueName: \"kubernetes.io/projected/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-kube-api-access-n6xkd\") pod \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.498800 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-config\") pod \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\" (UID: \"d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0\") " Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.499242 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0" (UID: "d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.499654 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-config" (OuterVolumeSpecName: "config") pod "d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0" (UID: "d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.505710 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-kube-api-access-n6xkd" (OuterVolumeSpecName: "kube-api-access-n6xkd") pod "d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0" (UID: "d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0"). InnerVolumeSpecName "kube-api-access-n6xkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.601034 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.601079 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6xkd\" (UniqueName: \"kubernetes.io/projected/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-kube-api-access-n6xkd\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.601094 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.604667 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.861320 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-wnfnj"] Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.879599 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.886809 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567c455747-wnfnj"] Dec 07 09:24:13 crc kubenswrapper[4838]: W1207 09:24:13.960605 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode06fb79a_09b7_42af_bb04_1ef9ba3f7560.slice/crio-699ed908185c78b2bcbf31f70c992340b49318e6ceb1e6e00d30f9ae7e45f8b1 WatchSource:0}: Error finding container 699ed908185c78b2bcbf31f70c992340b49318e6ceb1e6e00d30f9ae7e45f8b1: Status 404 returned error can't find the container with id 699ed908185c78b2bcbf31f70c992340b49318e6ceb1e6e00d30f9ae7e45f8b1 Dec 07 09:24:13 crc kubenswrapper[4838]: I1207 09:24:13.963328 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.014490 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-config\") pod \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.014775 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb5s5\" (UniqueName: \"kubernetes.io/projected/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-kube-api-access-mb5s5\") pod \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\" (UID: \"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a\") " Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.014936 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-config" (OuterVolumeSpecName: "config") pod "0256fa90-d1c3-4ef9-ab26-2e6fde11b77a" (UID: "0256fa90-d1c3-4ef9-ab26-2e6fde11b77a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.016552 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.109630 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-kube-api-access-mb5s5" (OuterVolumeSpecName: "kube-api-access-mb5s5") pod "0256fa90-d1c3-4ef9-ab26-2e6fde11b77a" (UID: "0256fa90-d1c3-4ef9-ab26-2e6fde11b77a"). InnerVolumeSpecName "kube-api-access-mb5s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.118334 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb5s5\" (UniqueName: \"kubernetes.io/projected/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a-kube-api-access-mb5s5\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.474580 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e06fb79a-09b7-42af-bb04-1ef9ba3f7560","Type":"ContainerStarted","Data":"699ed908185c78b2bcbf31f70c992340b49318e6ceb1e6e00d30f9ae7e45f8b1"} Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.476019 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be998839-55b0-48de-9903-197e3684a4cf","Type":"ContainerStarted","Data":"9d104a1501adf71063cf8a0f02f1b02023df463c279bef243677e22b2075a0e1"} Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.477985 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"76cc3a82-a620-435a-99f0-9a91297d2ee2","Type":"ContainerStarted","Data":"3db485b1f3aca38bdfda2551508506eebe129e9d611e7f72305d24d63fc44d7d"} Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.479443 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.482512 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eae80b29-4c64-403b-be38-7fd30b7ba66f","Type":"ContainerStarted","Data":"73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2"} Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.483991 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-68szs" event={"ID":"0256fa90-d1c3-4ef9-ab26-2e6fde11b77a","Type":"ContainerDied","Data":"3c25f3ec119bfc2a5727f8817a2f8e6ad13ff8ef8f99109610a2814a0b983f4d"} Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.484040 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-68szs" Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.499276 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.544632704 podStartE2EDuration="32.499260716s" podCreationTimestamp="2025-12-07 09:23:42 +0000 UTC" firstStartedPulling="2025-12-07 09:23:52.288357619 +0000 UTC m=+1048.995676636" lastFinishedPulling="2025-12-07 09:24:13.242985631 +0000 UTC m=+1069.950304648" observedRunningTime="2025-12-07 09:24:14.496659062 +0000 UTC m=+1071.203978079" watchObservedRunningTime="2025-12-07 09:24:14.499260716 +0000 UTC m=+1071.206579733" Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.568571 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-68szs"] Dec 07 09:24:14 crc kubenswrapper[4838]: I1207 09:24:14.574586 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-68szs"] Dec 07 09:24:15 crc kubenswrapper[4838]: I1207 09:24:15.011576 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gb65g"] Dec 07 09:24:15 crc kubenswrapper[4838]: I1207 09:24:15.623287 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0256fa90-d1c3-4ef9-ab26-2e6fde11b77a" path="/var/lib/kubelet/pods/0256fa90-d1c3-4ef9-ab26-2e6fde11b77a/volumes" Dec 07 09:24:15 crc kubenswrapper[4838]: I1207 09:24:15.623664 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0" path="/var/lib/kubelet/pods/d34e8e3d-9e9c-4029-b0e2-a5ad7a5990b0/volumes" Dec 07 09:24:19 crc kubenswrapper[4838]: I1207 09:24:19.518124 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gb65g" event={"ID":"dc745e13-8d6d-4e12-a733-5c081c0e4283","Type":"ContainerStarted","Data":"5d688164ac61a634ea69e8dc28a790a767ee7c05814fda8564f33fc60213fe8f"} Dec 07 09:24:22 crc kubenswrapper[4838]: I1207 09:24:22.544356 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cec74e7b-a2aa-46a0-b318-61decd213728","Type":"ContainerStarted","Data":"c2130c573c235007071f9eaf11f610f7ff20351e809000b90fad2653a7cc6630"} Dec 07 09:24:22 crc kubenswrapper[4838]: I1207 09:24:22.550270 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e06fb79a-09b7-42af-bb04-1ef9ba3f7560","Type":"ContainerStarted","Data":"07e22e5a6f8a2a2cdcdace76d82131912435013fc91505955d65e2065052af9c"} Dec 07 09:24:22 crc kubenswrapper[4838]: I1207 09:24:22.558260 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be998839-55b0-48de-9903-197e3684a4cf","Type":"ContainerStarted","Data":"9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef"} Dec 07 09:24:22 crc kubenswrapper[4838]: I1207 09:24:22.559211 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 07 09:24:22 crc kubenswrapper[4838]: I1207 09:24:22.588524 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=29.964723413 podStartE2EDuration="38.588504137s" podCreationTimestamp="2025-12-07 09:23:44 +0000 UTC" firstStartedPulling="2025-12-07 09:24:13.62710619 +0000 UTC m=+1070.334425207" lastFinishedPulling="2025-12-07 09:24:22.250886914 +0000 UTC m=+1078.958205931" observedRunningTime="2025-12-07 09:24:22.582656521 +0000 UTC m=+1079.289975538" watchObservedRunningTime="2025-12-07 09:24:22.588504137 +0000 UTC m=+1079.295823154" Dec 07 09:24:22 crc kubenswrapper[4838]: I1207 09:24:22.909266 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.574527 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t" event={"ID":"a3ae58f2-6cd2-4782-96b1-a88597b25b37","Type":"ContainerStarted","Data":"a6303be44836bdc3250566ea89059a5701710984c61cd4840073f7d5f27b65ed"} Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.575570 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-sbl7t" Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.578444 4838 generic.go:334] "Generic (PLEG): container finished" podID="dc745e13-8d6d-4e12-a733-5c081c0e4283" containerID="f107dad8d4c097fe33fb7bc3390b777d6428c1b6098418a6bda6309bb6f994aa" exitCode=0 Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.578497 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gb65g" event={"ID":"dc745e13-8d6d-4e12-a733-5c081c0e4283","Type":"ContainerDied","Data":"f107dad8d4c097fe33fb7bc3390b777d6428c1b6098418a6bda6309bb6f994aa"} Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.585035 4838 generic.go:334] "Generic (PLEG): container finished" podID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerID="cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525" exitCode=0 Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.585165 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" event={"ID":"280b54c8-5174-4b27-b41e-b7bd28b27111","Type":"ContainerDied","Data":"cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525"} Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.588631 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454","Type":"ContainerStarted","Data":"fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7"} Dec 07 09:24:23 crc kubenswrapper[4838]: I1207 09:24:23.604271 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-sbl7t" podStartSLOduration=27.170498439 podStartE2EDuration="35.604254163s" podCreationTimestamp="2025-12-07 09:23:48 +0000 UTC" firstStartedPulling="2025-12-07 09:24:13.127566954 +0000 UTC m=+1069.834885961" lastFinishedPulling="2025-12-07 09:24:21.561322678 +0000 UTC m=+1078.268641685" observedRunningTime="2025-12-07 09:24:23.593490506 +0000 UTC m=+1080.300809523" watchObservedRunningTime="2025-12-07 09:24:23.604254163 +0000 UTC m=+1080.311573180" Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.599838 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gb65g" event={"ID":"dc745e13-8d6d-4e12-a733-5c081c0e4283","Type":"ContainerStarted","Data":"8e030331bf8912db032a5626202d7d55d091bba48d26f10fed64420904d7b4f2"} Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.600208 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gb65g" event={"ID":"dc745e13-8d6d-4e12-a733-5c081c0e4283","Type":"ContainerStarted","Data":"4ca6c61615a5d544396a7b6eb75faff46eb2d31714cae4defc264a691c7cfd0a"} Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.600225 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.600240 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.603927 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" event={"ID":"280b54c8-5174-4b27-b41e-b7bd28b27111","Type":"ContainerStarted","Data":"24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1"} Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.604421 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.625891 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-gb65g" podStartSLOduration=33.818583703 podStartE2EDuration="36.625873385s" podCreationTimestamp="2025-12-07 09:23:48 +0000 UTC" firstStartedPulling="2025-12-07 09:24:19.283080561 +0000 UTC m=+1075.990399588" lastFinishedPulling="2025-12-07 09:24:22.090370253 +0000 UTC m=+1078.797689270" observedRunningTime="2025-12-07 09:24:24.623182058 +0000 UTC m=+1081.330501085" watchObservedRunningTime="2025-12-07 09:24:24.625873385 +0000 UTC m=+1081.333192402" Dec 07 09:24:24 crc kubenswrapper[4838]: I1207 09:24:24.677502 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" podStartSLOduration=3.189605377 podStartE2EDuration="46.677485075s" podCreationTimestamp="2025-12-07 09:23:38 +0000 UTC" firstStartedPulling="2025-12-07 09:23:39.537108679 +0000 UTC m=+1036.244427696" lastFinishedPulling="2025-12-07 09:24:23.024988377 +0000 UTC m=+1079.732307394" observedRunningTime="2025-12-07 09:24:24.671471633 +0000 UTC m=+1081.378790640" watchObservedRunningTime="2025-12-07 09:24:24.677485075 +0000 UTC m=+1081.384804092" Dec 07 09:24:27 crc kubenswrapper[4838]: I1207 09:24:27.627049 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cec74e7b-a2aa-46a0-b318-61decd213728","Type":"ContainerStarted","Data":"ee5e1963db63235e0e0f2889f0c29e6500ac05156ccea9e42329d55cae85abab"} Dec 07 09:24:27 crc kubenswrapper[4838]: I1207 09:24:27.630059 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e06fb79a-09b7-42af-bb04-1ef9ba3f7560","Type":"ContainerStarted","Data":"832406694e125cbec414790d25b98386d0cb19e070805270aa5014875b407706"} Dec 07 09:24:27 crc kubenswrapper[4838]: I1207 09:24:27.630662 4838 generic.go:334] "Generic (PLEG): container finished" podID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerID="0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d" exitCode=0 Dec 07 09:24:27 crc kubenswrapper[4838]: I1207 09:24:27.630751 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" event={"ID":"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e","Type":"ContainerDied","Data":"0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d"} Dec 07 09:24:27 crc kubenswrapper[4838]: I1207 09:24:27.634087 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d111910-1e1a-4b37-87c2-5c8f55110a73","Type":"ContainerStarted","Data":"3065e7e0a00e33b3977117f97c877c2a21324987118a888369e57c67af453c68"} Dec 07 09:24:27 crc kubenswrapper[4838]: I1207 09:24:27.720882 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=25.48643407 podStartE2EDuration="38.720863689s" podCreationTimestamp="2025-12-07 09:23:49 +0000 UTC" firstStartedPulling="2025-12-07 09:24:13.127590515 +0000 UTC m=+1069.834909532" lastFinishedPulling="2025-12-07 09:24:26.362020134 +0000 UTC m=+1083.069339151" observedRunningTime="2025-12-07 09:24:27.703994919 +0000 UTC m=+1084.411313956" watchObservedRunningTime="2025-12-07 09:24:27.720863689 +0000 UTC m=+1084.428182696" Dec 07 09:24:27 crc kubenswrapper[4838]: I1207 09:24:27.769634 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=25.337512846 podStartE2EDuration="37.769610527s" podCreationTimestamp="2025-12-07 09:23:50 +0000 UTC" firstStartedPulling="2025-12-07 09:24:13.963691854 +0000 UTC m=+1070.671010871" lastFinishedPulling="2025-12-07 09:24:26.395789535 +0000 UTC m=+1083.103108552" observedRunningTime="2025-12-07 09:24:27.76864728 +0000 UTC m=+1084.475966297" watchObservedRunningTime="2025-12-07 09:24:27.769610527 +0000 UTC m=+1084.476929554" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.225889 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.267785 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.641845 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c","Type":"ContainerStarted","Data":"56d38aafc7235cb6646a720353b4d3c5512888f1e3bc9ad867762a2b766c27b8"} Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.645262 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" event={"ID":"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e","Type":"ContainerStarted","Data":"dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14"} Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.645707 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.694979 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" podStartSLOduration=-9223371985.159824 podStartE2EDuration="51.694952388s" podCreationTimestamp="2025-12-07 09:23:37 +0000 UTC" firstStartedPulling="2025-12-07 09:23:39.218023403 +0000 UTC m=+1035.925342420" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:24:28.685065636 +0000 UTC m=+1085.392384653" watchObservedRunningTime="2025-12-07 09:24:28.694952388 +0000 UTC m=+1085.402271405" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.695243 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.734095 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.833836 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-b6n8d"] Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.951055 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-xwxqx"] Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.952639 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.955378 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 07 09:24:28 crc kubenswrapper[4838]: I1207 09:24:28.992206 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-xwxqx"] Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.096526 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-config\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.096582 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt6d8\" (UniqueName: \"kubernetes.io/projected/b1d57e03-d637-48a2-9142-614282a01260-kube-api-access-lt6d8\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.096627 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-dns-svc\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.096717 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.142837 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-j7t4b"] Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.143977 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.146274 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.170317 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j7t4b"] Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.197867 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-config\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.197936 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt6d8\" (UniqueName: \"kubernetes.io/projected/b1d57e03-d637-48a2-9142-614282a01260-kube-api-access-lt6d8\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.197971 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-dns-svc\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.198025 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.198757 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-config\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.198839 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.198969 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-dns-svc\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.226842 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt6d8\" (UniqueName: \"kubernetes.io/projected/b1d57e03-d637-48a2-9142-614282a01260-kube-api-access-lt6d8\") pod \"dnsmasq-dns-7cb7d9565c-xwxqx\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.266750 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.299676 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7b95347-0585-4016-a3fa-a17c6d9042a3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.300039 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e7b95347-0585-4016-a3fa-a17c6d9042a3-ovs-rundir\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.300086 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e7b95347-0585-4016-a3fa-a17c6d9042a3-ovn-rundir\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.300107 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7b95347-0585-4016-a3fa-a17c6d9042a3-combined-ca-bundle\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.300161 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8prp\" (UniqueName: \"kubernetes.io/projected/e7b95347-0585-4016-a3fa-a17c6d9042a3-kube-api-access-z8prp\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.300181 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7b95347-0585-4016-a3fa-a17c6d9042a3-config\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.403603 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8prp\" (UniqueName: \"kubernetes.io/projected/e7b95347-0585-4016-a3fa-a17c6d9042a3-kube-api-access-z8prp\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.403653 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7b95347-0585-4016-a3fa-a17c6d9042a3-config\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.403709 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7b95347-0585-4016-a3fa-a17c6d9042a3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.403732 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e7b95347-0585-4016-a3fa-a17c6d9042a3-ovs-rundir\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.403769 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e7b95347-0585-4016-a3fa-a17c6d9042a3-ovn-rundir\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.403788 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7b95347-0585-4016-a3fa-a17c6d9042a3-combined-ca-bundle\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.405432 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e7b95347-0585-4016-a3fa-a17c6d9042a3-ovs-rundir\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.406751 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e7b95347-0585-4016-a3fa-a17c6d9042a3-ovn-rundir\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.407579 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7b95347-0585-4016-a3fa-a17c6d9042a3-config\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.412743 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7b95347-0585-4016-a3fa-a17c6d9042a3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.415342 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7b95347-0585-4016-a3fa-a17c6d9042a3-combined-ca-bundle\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.440524 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8prp\" (UniqueName: \"kubernetes.io/projected/e7b95347-0585-4016-a3fa-a17c6d9042a3-kube-api-access-z8prp\") pod \"ovn-controller-metrics-j7t4b\" (UID: \"e7b95347-0585-4016-a3fa-a17c6d9042a3\") " pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.459030 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j7t4b" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.565392 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-xwxqx"] Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.651173 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-fm8dz"] Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.683239 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerName="dnsmasq-dns" containerID="cri-o://dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14" gracePeriod=10 Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.692924 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-fm8dz"] Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.693632 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.693071 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.701405 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.814289 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.814348 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-config\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.815382 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.815548 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-dns-svc\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.815629 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d85gt\" (UniqueName: \"kubernetes.io/projected/3dcec780-852f-4a5a-b977-a92fd070416c-kube-api-access-d85gt\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.918588 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.918665 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-dns-svc\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.918703 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d85gt\" (UniqueName: \"kubernetes.io/projected/3dcec780-852f-4a5a-b977-a92fd070416c-kube-api-access-d85gt\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.918768 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.918786 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-config\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.919503 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.919598 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-dns-svc\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.919785 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-config\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.920029 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.952404 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d85gt\" (UniqueName: \"kubernetes.io/projected/3dcec780-852f-4a5a-b977-a92fd070416c-kube-api-access-d85gt\") pod \"dnsmasq-dns-db7757ddc-fm8dz\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:29 crc kubenswrapper[4838]: I1207 09:24:29.980586 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-xwxqx"] Dec 07 09:24:30 crc kubenswrapper[4838]: W1207 09:24:29.998868 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1d57e03_d637_48a2_9142_614282a01260.slice/crio-867e2499a70c8f0e907e5977ebbb58c14c043e925b522928eff04a50d6af8b44 WatchSource:0}: Error finding container 867e2499a70c8f0e907e5977ebbb58c14c043e925b522928eff04a50d6af8b44: Status 404 returned error can't find the container with id 867e2499a70c8f0e907e5977ebbb58c14c043e925b522928eff04a50d6af8b44 Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.044111 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.066372 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.086362 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.109153 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.162163 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j7t4b"] Dec 07 09:24:30 crc kubenswrapper[4838]: W1207 09:24:30.171917 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7b95347_0585_4016_a3fa_a17c6d9042a3.slice/crio-af030edeb1c69d26ee254e7e4489fd4f44d67e40ac8ab22edcb37057ba8dd972 WatchSource:0}: Error finding container af030edeb1c69d26ee254e7e4489fd4f44d67e40ac8ab22edcb37057ba8dd972: Status 404 returned error can't find the container with id af030edeb1c69d26ee254e7e4489fd4f44d67e40ac8ab22edcb37057ba8dd972 Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.225626 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-dns-svc\") pod \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.225711 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-config\") pod \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.225851 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlvv9\" (UniqueName: \"kubernetes.io/projected/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-kube-api-access-tlvv9\") pod \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\" (UID: \"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e\") " Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.229791 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-kube-api-access-tlvv9" (OuterVolumeSpecName: "kube-api-access-tlvv9") pod "0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" (UID: "0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e"). InnerVolumeSpecName "kube-api-access-tlvv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.267398 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-config" (OuterVolumeSpecName: "config") pod "0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" (UID: "0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.290947 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" (UID: "0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.328616 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlvv9\" (UniqueName: \"kubernetes.io/projected/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-kube-api-access-tlvv9\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.328887 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.328904 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.580427 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-fm8dz"] Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.692394 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" event={"ID":"3dcec780-852f-4a5a-b977-a92fd070416c","Type":"ContainerStarted","Data":"07bee7c5811e74f8d67dc5ab4f9afa50e2f69f99b92fe1fcd9e941e4fc53629b"} Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.694312 4838 generic.go:334] "Generic (PLEG): container finished" podID="b1d57e03-d637-48a2-9142-614282a01260" containerID="38e8a4c442b6f055ca6a6613488e6135619a14bba4da7f1077ccaebdc4f4f59f" exitCode=0 Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.694435 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" event={"ID":"b1d57e03-d637-48a2-9142-614282a01260","Type":"ContainerDied","Data":"38e8a4c442b6f055ca6a6613488e6135619a14bba4da7f1077ccaebdc4f4f59f"} Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.694744 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" event={"ID":"b1d57e03-d637-48a2-9142-614282a01260","Type":"ContainerStarted","Data":"867e2499a70c8f0e907e5977ebbb58c14c043e925b522928eff04a50d6af8b44"} Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.697052 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j7t4b" event={"ID":"e7b95347-0585-4016-a3fa-a17c6d9042a3","Type":"ContainerStarted","Data":"cbcb849f88397be14527e879104615455a39e7fd17336a9d0ad57175f1316ffa"} Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.697090 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j7t4b" event={"ID":"e7b95347-0585-4016-a3fa-a17c6d9042a3","Type":"ContainerStarted","Data":"af030edeb1c69d26ee254e7e4489fd4f44d67e40ac8ab22edcb37057ba8dd972"} Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.700593 4838 generic.go:334] "Generic (PLEG): container finished" podID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerID="dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14" exitCode=0 Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.700986 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" event={"ID":"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e","Type":"ContainerDied","Data":"dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14"} Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.701015 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" event={"ID":"0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e","Type":"ContainerDied","Data":"1603d085eb920214d08b0edb5a43658cbad30737fc98e958a96279e673453e7e"} Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.701031 4838 scope.go:117] "RemoveContainer" containerID="dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.701280 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-b6n8d" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.701399 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.747955 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-j7t4b" podStartSLOduration=1.747934919 podStartE2EDuration="1.747934919s" podCreationTimestamp="2025-12-07 09:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:24:30.744573863 +0000 UTC m=+1087.451892890" watchObservedRunningTime="2025-12-07 09:24:30.747934919 +0000 UTC m=+1087.455253936" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.768149 4838 scope.go:117] "RemoveContainer" containerID="0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.769138 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-b6n8d"] Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.778517 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-b6n8d"] Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.786153 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.807734 4838 scope.go:117] "RemoveContainer" containerID="dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14" Dec 07 09:24:30 crc kubenswrapper[4838]: E1207 09:24:30.815921 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14\": container with ID starting with dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14 not found: ID does not exist" containerID="dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.815966 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14"} err="failed to get container status \"dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14\": rpc error: code = NotFound desc = could not find container \"dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14\": container with ID starting with dc232259817b2d1ab0a95053613472aec7c137e2afd384093bcc9fdb8a169b14 not found: ID does not exist" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.815992 4838 scope.go:117] "RemoveContainer" containerID="0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d" Dec 07 09:24:30 crc kubenswrapper[4838]: E1207 09:24:30.817293 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d\": container with ID starting with 0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d not found: ID does not exist" containerID="0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d" Dec 07 09:24:30 crc kubenswrapper[4838]: I1207 09:24:30.817319 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d"} err="failed to get container status \"0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d\": rpc error: code = NotFound desc = could not find container \"0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d\": container with ID starting with 0923bd280c4059cf8138d92a94c45e408872ce0fbec12740c45763eb0797155d not found: ID does not exist" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.102955 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 07 09:24:31 crc kubenswrapper[4838]: E1207 09:24:31.103611 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerName="dnsmasq-dns" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.103629 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerName="dnsmasq-dns" Dec 07 09:24:31 crc kubenswrapper[4838]: E1207 09:24:31.103653 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerName="init" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.103660 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerName="init" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.103827 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" containerName="dnsmasq-dns" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.104601 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.107241 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.107374 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.107420 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-22gzk" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.110075 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.110505 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.127830 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245345 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-config\") pod \"b1d57e03-d637-48a2-9142-614282a01260\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245443 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-dns-svc\") pod \"b1d57e03-d637-48a2-9142-614282a01260\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245510 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt6d8\" (UniqueName: \"kubernetes.io/projected/b1d57e03-d637-48a2-9142-614282a01260-kube-api-access-lt6d8\") pod \"b1d57e03-d637-48a2-9142-614282a01260\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245534 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-ovsdbserver-sb\") pod \"b1d57e03-d637-48a2-9142-614282a01260\" (UID: \"b1d57e03-d637-48a2-9142-614282a01260\") " Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245733 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-scripts\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245844 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245886 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245922 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245943 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkcfk\" (UniqueName: \"kubernetes.io/projected/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-kube-api-access-zkcfk\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.245974 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.246026 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-config\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.250181 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d57e03-d637-48a2-9142-614282a01260-kube-api-access-lt6d8" (OuterVolumeSpecName: "kube-api-access-lt6d8") pod "b1d57e03-d637-48a2-9142-614282a01260" (UID: "b1d57e03-d637-48a2-9142-614282a01260"). InnerVolumeSpecName "kube-api-access-lt6d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.264667 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1d57e03-d637-48a2-9142-614282a01260" (UID: "b1d57e03-d637-48a2-9142-614282a01260"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.271389 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b1d57e03-d637-48a2-9142-614282a01260" (UID: "b1d57e03-d637-48a2-9142-614282a01260"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.272225 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-config" (OuterVolumeSpecName: "config") pod "b1d57e03-d637-48a2-9142-614282a01260" (UID: "b1d57e03-d637-48a2-9142-614282a01260"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.347996 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348095 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348141 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348168 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkcfk\" (UniqueName: \"kubernetes.io/projected/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-kube-api-access-zkcfk\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348206 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348262 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-config\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348300 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-scripts\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348409 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt6d8\" (UniqueName: \"kubernetes.io/projected/b1d57e03-d637-48a2-9142-614282a01260-kube-api-access-lt6d8\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348426 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348439 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.348451 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1d57e03-d637-48a2-9142-614282a01260-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.349272 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.349619 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-scripts\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.349760 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-config\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.355221 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.355876 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.356620 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.367756 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkcfk\" (UniqueName: \"kubernetes.io/projected/065bed0f-e9fc-4d9a-a455-46d9d77e7bb2-kube-api-access-zkcfk\") pod \"ovn-northd-0\" (UID: \"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2\") " pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.440786 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.624925 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e" path="/var/lib/kubelet/pods/0d37f7b3-d058-4f8f-8fb4-f7e6e0da552e/volumes" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.734543 4838 generic.go:334] "Generic (PLEG): container finished" podID="3d111910-1e1a-4b37-87c2-5c8f55110a73" containerID="3065e7e0a00e33b3977117f97c877c2a21324987118a888369e57c67af453c68" exitCode=0 Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.734600 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d111910-1e1a-4b37-87c2-5c8f55110a73","Type":"ContainerDied","Data":"3065e7e0a00e33b3977117f97c877c2a21324987118a888369e57c67af453c68"} Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.745031 4838 generic.go:334] "Generic (PLEG): container finished" podID="3dcec780-852f-4a5a-b977-a92fd070416c" containerID="5eb7b29805ce8fb4b7990c6b189e9f9990dada903daeac5e7f823d8a21266dbe" exitCode=0 Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.745089 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" event={"ID":"3dcec780-852f-4a5a-b977-a92fd070416c","Type":"ContainerDied","Data":"5eb7b29805ce8fb4b7990c6b189e9f9990dada903daeac5e7f823d8a21266dbe"} Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.763950 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.764531 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-xwxqx" event={"ID":"b1d57e03-d637-48a2-9142-614282a01260","Type":"ContainerDied","Data":"867e2499a70c8f0e907e5977ebbb58c14c043e925b522928eff04a50d6af8b44"} Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.764584 4838 scope.go:117] "RemoveContainer" containerID="38e8a4c442b6f055ca6a6613488e6135619a14bba4da7f1077ccaebdc4f4f59f" Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.868328 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 07 09:24:31 crc kubenswrapper[4838]: I1207 09:24:31.976908 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-xwxqx"] Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.002600 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-xwxqx"] Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.772082 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2","Type":"ContainerStarted","Data":"88ce26aba478b9cd2d022aaac2e0d2808301f8f7bd0635919b2fd8900150892c"} Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.776915 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d111910-1e1a-4b37-87c2-5c8f55110a73","Type":"ContainerStarted","Data":"badcdd379c5b551c19e6a412fdb1dc07453ab4fa199399f6a520395ae1bd0b14"} Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.785790 4838 generic.go:334] "Generic (PLEG): container finished" podID="898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c" containerID="56d38aafc7235cb6646a720353b4d3c5512888f1e3bc9ad867762a2b766c27b8" exitCode=0 Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.785897 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c","Type":"ContainerDied","Data":"56d38aafc7235cb6646a720353b4d3c5512888f1e3bc9ad867762a2b766c27b8"} Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.790291 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" event={"ID":"3dcec780-852f-4a5a-b977-a92fd070416c","Type":"ContainerStarted","Data":"9e50b3fe4418a1ee0920041fb831a5a2ad19591f0c794ec8b49af11dbed1720b"} Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.790503 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.804560 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.46449703 podStartE2EDuration="53.804545235s" podCreationTimestamp="2025-12-07 09:23:39 +0000 UTC" firstStartedPulling="2025-12-07 09:23:42.057090079 +0000 UTC m=+1038.764409096" lastFinishedPulling="2025-12-07 09:24:26.397138294 +0000 UTC m=+1083.104457301" observedRunningTime="2025-12-07 09:24:32.798498232 +0000 UTC m=+1089.505817249" watchObservedRunningTime="2025-12-07 09:24:32.804545235 +0000 UTC m=+1089.511864252" Dec 07 09:24:32 crc kubenswrapper[4838]: I1207 09:24:32.843372 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" podStartSLOduration=3.84335617 podStartE2EDuration="3.84335617s" podCreationTimestamp="2025-12-07 09:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:24:32.837477082 +0000 UTC m=+1089.544796099" watchObservedRunningTime="2025-12-07 09:24:32.84335617 +0000 UTC m=+1089.550675187" Dec 07 09:24:33 crc kubenswrapper[4838]: I1207 09:24:33.624536 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1d57e03-d637-48a2-9142-614282a01260" path="/var/lib/kubelet/pods/b1d57e03-d637-48a2-9142-614282a01260/volumes" Dec 07 09:24:33 crc kubenswrapper[4838]: I1207 09:24:33.797126 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2","Type":"ContainerStarted","Data":"6260ad3456726bd4311a72f75732e5cfb08a6ca886528a6230c54525f53a23dc"} Dec 07 09:24:33 crc kubenswrapper[4838]: I1207 09:24:33.797184 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"065bed0f-e9fc-4d9a-a455-46d9d77e7bb2","Type":"ContainerStarted","Data":"19da68d3031aed718c459321559b64f5d09315f1627345f7b667bf7ceabae9c4"} Dec 07 09:24:33 crc kubenswrapper[4838]: I1207 09:24:33.797319 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 07 09:24:33 crc kubenswrapper[4838]: I1207 09:24:33.799031 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c","Type":"ContainerStarted","Data":"87ae1d0b707678aeecb3b3705b1dd6db9bdaf21fe35daaecc223b20b0c58c258"} Dec 07 09:24:33 crc kubenswrapper[4838]: I1207 09:24:33.817762 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.608973204 podStartE2EDuration="2.817743046s" podCreationTimestamp="2025-12-07 09:24:31 +0000 UTC" firstStartedPulling="2025-12-07 09:24:31.881435977 +0000 UTC m=+1088.588754994" lastFinishedPulling="2025-12-07 09:24:33.090205819 +0000 UTC m=+1089.797524836" observedRunningTime="2025-12-07 09:24:33.815921824 +0000 UTC m=+1090.523240851" watchObservedRunningTime="2025-12-07 09:24:33.817743046 +0000 UTC m=+1090.525062063" Dec 07 09:24:33 crc kubenswrapper[4838]: I1207 09:24:33.844648 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371984.010143 podStartE2EDuration="52.844631622s" podCreationTimestamp="2025-12-07 09:23:41 +0000 UTC" firstStartedPulling="2025-12-07 09:23:43.14822134 +0000 UTC m=+1039.855540357" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:24:33.84140748 +0000 UTC m=+1090.548726507" watchObservedRunningTime="2025-12-07 09:24:33.844631622 +0000 UTC m=+1090.551950639" Dec 07 09:24:34 crc kubenswrapper[4838]: E1207 09:24:34.433338 4838 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.73:33308->38.102.83.73:33371: write tcp 38.102.83.73:33308->38.102.83.73:33371: write: broken pipe Dec 07 09:24:34 crc kubenswrapper[4838]: I1207 09:24:34.876224 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.068018 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.139571 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-sgg8c"] Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.140292 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerName="dnsmasq-dns" containerID="cri-o://24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1" gracePeriod=10 Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.597456 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.669490 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-config\") pod \"280b54c8-5174-4b27-b41e-b7bd28b27111\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.669575 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7pdp\" (UniqueName: \"kubernetes.io/projected/280b54c8-5174-4b27-b41e-b7bd28b27111-kube-api-access-j7pdp\") pod \"280b54c8-5174-4b27-b41e-b7bd28b27111\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.669683 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-dns-svc\") pod \"280b54c8-5174-4b27-b41e-b7bd28b27111\" (UID: \"280b54c8-5174-4b27-b41e-b7bd28b27111\") " Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.680435 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/280b54c8-5174-4b27-b41e-b7bd28b27111-kube-api-access-j7pdp" (OuterVolumeSpecName: "kube-api-access-j7pdp") pod "280b54c8-5174-4b27-b41e-b7bd28b27111" (UID: "280b54c8-5174-4b27-b41e-b7bd28b27111"). InnerVolumeSpecName "kube-api-access-j7pdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.714361 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-config" (OuterVolumeSpecName: "config") pod "280b54c8-5174-4b27-b41e-b7bd28b27111" (UID: "280b54c8-5174-4b27-b41e-b7bd28b27111"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.740866 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "280b54c8-5174-4b27-b41e-b7bd28b27111" (UID: "280b54c8-5174-4b27-b41e-b7bd28b27111"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.772252 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.772332 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7pdp\" (UniqueName: \"kubernetes.io/projected/280b54c8-5174-4b27-b41e-b7bd28b27111-kube-api-access-j7pdp\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.772352 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/280b54c8-5174-4b27-b41e-b7bd28b27111-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.850874 4838 generic.go:334] "Generic (PLEG): container finished" podID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerID="24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1" exitCode=0 Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.850946 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" event={"ID":"280b54c8-5174-4b27-b41e-b7bd28b27111","Type":"ContainerDied","Data":"24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1"} Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.850980 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" event={"ID":"280b54c8-5174-4b27-b41e-b7bd28b27111","Type":"ContainerDied","Data":"4f9420f4efe2bc8cd64d47373180ac374e82e9fd3feb11d90eb706e162bd9b9c"} Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.851058 4838 scope.go:117] "RemoveContainer" containerID="24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.851288 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-sgg8c" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.870419 4838 scope.go:117] "RemoveContainer" containerID="cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.886845 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-sgg8c"] Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.894469 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-sgg8c"] Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.901883 4838 scope.go:117] "RemoveContainer" containerID="24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1" Dec 07 09:24:40 crc kubenswrapper[4838]: E1207 09:24:40.902782 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1\": container with ID starting with 24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1 not found: ID does not exist" containerID="24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.902855 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1"} err="failed to get container status \"24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1\": rpc error: code = NotFound desc = could not find container \"24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1\": container with ID starting with 24855e2847d2da496f32d75f0b1aae9cff44a617b3aa67095b126339e11469a1 not found: ID does not exist" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.902882 4838 scope.go:117] "RemoveContainer" containerID="cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525" Dec 07 09:24:40 crc kubenswrapper[4838]: E1207 09:24:40.903296 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525\": container with ID starting with cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525 not found: ID does not exist" containerID="cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525" Dec 07 09:24:40 crc kubenswrapper[4838]: I1207 09:24:40.903321 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525"} err="failed to get container status \"cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525\": rpc error: code = NotFound desc = could not find container \"cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525\": container with ID starting with cc83490ba6b314946b83c20e01a8c3d573886c86bede6edb68707a60c991b525 not found: ID does not exist" Dec 07 09:24:41 crc kubenswrapper[4838]: I1207 09:24:41.132585 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 07 09:24:41 crc kubenswrapper[4838]: I1207 09:24:41.132705 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 07 09:24:41 crc kubenswrapper[4838]: I1207 09:24:41.212331 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 07 09:24:41 crc kubenswrapper[4838]: I1207 09:24:41.623514 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" path="/var/lib/kubelet/pods/280b54c8-5174-4b27-b41e-b7bd28b27111/volumes" Dec 07 09:24:41 crc kubenswrapper[4838]: I1207 09:24:41.947668 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.457258 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.458152 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.520370 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ecf1-account-create-update-k27fv"] Dec 07 09:24:42 crc kubenswrapper[4838]: E1207 09:24:42.520777 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d57e03-d637-48a2-9142-614282a01260" containerName="init" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.520798 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d57e03-d637-48a2-9142-614282a01260" containerName="init" Dec 07 09:24:42 crc kubenswrapper[4838]: E1207 09:24:42.520843 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerName="init" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.520852 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerName="init" Dec 07 09:24:42 crc kubenswrapper[4838]: E1207 09:24:42.520870 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerName="dnsmasq-dns" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.520878 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerName="dnsmasq-dns" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.521058 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="280b54c8-5174-4b27-b41e-b7bd28b27111" containerName="dnsmasq-dns" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.521074 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d57e03-d637-48a2-9142-614282a01260" containerName="init" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.521723 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.523617 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.544461 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ecf1-account-create-update-k27fv"] Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.550058 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g9b52"] Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.550976 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.567947 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g9b52"] Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.605096 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-operator-scripts\") pod \"keystone-db-create-g9b52\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.605427 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd2vq\" (UniqueName: \"kubernetes.io/projected/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-kube-api-access-pd2vq\") pod \"keystone-db-create-g9b52\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.605454 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwdz\" (UniqueName: \"kubernetes.io/projected/298292bd-8186-4646-ad35-8058c9963946-kube-api-access-jpwdz\") pod \"keystone-ecf1-account-create-update-k27fv\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.605484 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298292bd-8186-4646-ad35-8058c9963946-operator-scripts\") pod \"keystone-ecf1-account-create-update-k27fv\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.610060 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.706643 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd2vq\" (UniqueName: \"kubernetes.io/projected/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-kube-api-access-pd2vq\") pod \"keystone-db-create-g9b52\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.706694 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwdz\" (UniqueName: \"kubernetes.io/projected/298292bd-8186-4646-ad35-8058c9963946-kube-api-access-jpwdz\") pod \"keystone-ecf1-account-create-update-k27fv\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.706730 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298292bd-8186-4646-ad35-8058c9963946-operator-scripts\") pod \"keystone-ecf1-account-create-update-k27fv\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.706792 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-operator-scripts\") pod \"keystone-db-create-g9b52\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.708168 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-operator-scripts\") pod \"keystone-db-create-g9b52\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.708317 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298292bd-8186-4646-ad35-8058c9963946-operator-scripts\") pod \"keystone-ecf1-account-create-update-k27fv\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.728033 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwdz\" (UniqueName: \"kubernetes.io/projected/298292bd-8186-4646-ad35-8058c9963946-kube-api-access-jpwdz\") pod \"keystone-ecf1-account-create-update-k27fv\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.739693 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd2vq\" (UniqueName: \"kubernetes.io/projected/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-kube-api-access-pd2vq\") pod \"keystone-db-create-g9b52\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.846602 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.889184 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5jhzd"] Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.890204 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.905560 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5jhzd"] Dec 07 09:24:42 crc kubenswrapper[4838]: I1207 09:24:42.919260 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:42.996561 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-21a4-account-create-update-p44qz"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:42.998985 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.005449 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.014285 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msmhg\" (UniqueName: \"kubernetes.io/projected/6863df38-7011-4203-adb9-117671f13bf4-kube-api-access-msmhg\") pod \"placement-db-create-5jhzd\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.014362 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6863df38-7011-4203-adb9-117671f13bf4-operator-scripts\") pod \"placement-db-create-5jhzd\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.029359 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-21a4-account-create-update-p44qz"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.069761 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.096166 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5nft8"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.097282 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.116364 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msmhg\" (UniqueName: \"kubernetes.io/projected/6863df38-7011-4203-adb9-117671f13bf4-kube-api-access-msmhg\") pod \"placement-db-create-5jhzd\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.116419 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8szk\" (UniqueName: \"kubernetes.io/projected/05852af0-5b52-4356-97e5-2fb3e9800c65-kube-api-access-z8szk\") pod \"placement-21a4-account-create-update-p44qz\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.116449 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6369855a-6de3-482c-8b9c-3e0263b7979f-operator-scripts\") pod \"glance-db-create-5nft8\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.116470 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6863df38-7011-4203-adb9-117671f13bf4-operator-scripts\") pod \"placement-db-create-5jhzd\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.116488 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffxk5\" (UniqueName: \"kubernetes.io/projected/6369855a-6de3-482c-8b9c-3e0263b7979f-kube-api-access-ffxk5\") pod \"glance-db-create-5nft8\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.116512 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05852af0-5b52-4356-97e5-2fb3e9800c65-operator-scripts\") pod \"placement-21a4-account-create-update-p44qz\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.117323 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6863df38-7011-4203-adb9-117671f13bf4-operator-scripts\") pod \"placement-db-create-5jhzd\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.136191 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msmhg\" (UniqueName: \"kubernetes.io/projected/6863df38-7011-4203-adb9-117671f13bf4-kube-api-access-msmhg\") pod \"placement-db-create-5jhzd\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.157062 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5nft8"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.202056 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-1db7-account-create-update-6m8pz"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.203115 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.209273 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.213317 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1db7-account-create-update-6m8pz"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.219607 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05852af0-5b52-4356-97e5-2fb3e9800c65-operator-scripts\") pod \"placement-21a4-account-create-update-p44qz\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.219655 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82kbx\" (UniqueName: \"kubernetes.io/projected/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-kube-api-access-82kbx\") pod \"glance-1db7-account-create-update-6m8pz\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.219713 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-operator-scripts\") pod \"glance-1db7-account-create-update-6m8pz\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.219790 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8szk\" (UniqueName: \"kubernetes.io/projected/05852af0-5b52-4356-97e5-2fb3e9800c65-kube-api-access-z8szk\") pod \"placement-21a4-account-create-update-p44qz\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.219828 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6369855a-6de3-482c-8b9c-3e0263b7979f-operator-scripts\") pod \"glance-db-create-5nft8\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.219847 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffxk5\" (UniqueName: \"kubernetes.io/projected/6369855a-6de3-482c-8b9c-3e0263b7979f-kube-api-access-ffxk5\") pod \"glance-db-create-5nft8\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.220299 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05852af0-5b52-4356-97e5-2fb3e9800c65-operator-scripts\") pod \"placement-21a4-account-create-update-p44qz\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.220645 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6369855a-6de3-482c-8b9c-3e0263b7979f-operator-scripts\") pod \"glance-db-create-5nft8\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.249115 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffxk5\" (UniqueName: \"kubernetes.io/projected/6369855a-6de3-482c-8b9c-3e0263b7979f-kube-api-access-ffxk5\") pod \"glance-db-create-5nft8\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.257298 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8szk\" (UniqueName: \"kubernetes.io/projected/05852af0-5b52-4356-97e5-2fb3e9800c65-kube-api-access-z8szk\") pod \"placement-21a4-account-create-update-p44qz\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.308111 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.321903 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82kbx\" (UniqueName: \"kubernetes.io/projected/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-kube-api-access-82kbx\") pod \"glance-1db7-account-create-update-6m8pz\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.323585 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-operator-scripts\") pod \"glance-1db7-account-create-update-6m8pz\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.324736 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-operator-scripts\") pod \"glance-1db7-account-create-update-6m8pz\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.331286 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.340638 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82kbx\" (UniqueName: \"kubernetes.io/projected/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-kube-api-access-82kbx\") pod \"glance-1db7-account-create-update-6m8pz\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.427439 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ecf1-account-create-update-k27fv"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.435437 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5nft8" Dec 07 09:24:43 crc kubenswrapper[4838]: W1207 09:24:43.443192 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod298292bd_8186_4646_ad35_8058c9963946.slice/crio-7d1f5716edabe5d7d8c7fbda787917078077c31d03aa40086816c2272e11062d WatchSource:0}: Error finding container 7d1f5716edabe5d7d8c7fbda787917078077c31d03aa40086816c2272e11062d: Status 404 returned error can't find the container with id 7d1f5716edabe5d7d8c7fbda787917078077c31d03aa40086816c2272e11062d Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.537563 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.555682 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g9b52"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.740315 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5jhzd"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.786361 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-21a4-account-create-update-p44qz"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.900858 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ecf1-account-create-update-k27fv" event={"ID":"298292bd-8186-4646-ad35-8058c9963946","Type":"ContainerStarted","Data":"2ba8fddc8a33dfaed82c785f863267073eb9608fab8ab17f0c11f872b5c6e6d8"} Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.900902 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ecf1-account-create-update-k27fv" event={"ID":"298292bd-8186-4646-ad35-8058c9963946","Type":"ContainerStarted","Data":"7d1f5716edabe5d7d8c7fbda787917078077c31d03aa40086816c2272e11062d"} Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.902142 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5nft8"] Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.924022 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-21a4-account-create-update-p44qz" event={"ID":"05852af0-5b52-4356-97e5-2fb3e9800c65","Type":"ContainerStarted","Data":"3e81fc388eb254e8fb282ba5ab3b5907ae05a3e198e8abe7bb1ff193660d35da"} Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.937543 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-ecf1-account-create-update-k27fv" podStartSLOduration=1.937524132 podStartE2EDuration="1.937524132s" podCreationTimestamp="2025-12-07 09:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:24:43.928794413 +0000 UTC m=+1100.636113430" watchObservedRunningTime="2025-12-07 09:24:43.937524132 +0000 UTC m=+1100.644843169" Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.961095 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g9b52" event={"ID":"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e","Type":"ContainerStarted","Data":"3dfe2c1926797d8cb64d8a414bc43c6b7980f204a0e84e8ebeab75650335e06b"} Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.961142 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g9b52" event={"ID":"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e","Type":"ContainerStarted","Data":"011ebbfca5ecd24e78fdfa529461635975c9ef1f032756a50e1ab7965ecd7303"} Dec 07 09:24:43 crc kubenswrapper[4838]: I1207 09:24:43.996398 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5jhzd" event={"ID":"6863df38-7011-4203-adb9-117671f13bf4","Type":"ContainerStarted","Data":"01cfc8331c9c0f91485029bce4984a84afa9946b538ef3fa44287f78ba0fe899"} Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.250009 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-g9b52" podStartSLOduration=2.249992199 podStartE2EDuration="2.249992199s" podCreationTimestamp="2025-12-07 09:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:24:43.996658725 +0000 UTC m=+1100.703977742" watchObservedRunningTime="2025-12-07 09:24:44.249992199 +0000 UTC m=+1100.957311216" Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.251218 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1db7-account-create-update-6m8pz"] Dec 07 09:24:44 crc kubenswrapper[4838]: W1207 09:24:44.260488 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod305e31d7_29e9_44f4_9dec_2aa2fd20c46c.slice/crio-b6ee9a80466069fa1e166317d123a0c4bf497b3097a1399069ca44a9effb32fd WatchSource:0}: Error finding container b6ee9a80466069fa1e166317d123a0c4bf497b3097a1399069ca44a9effb32fd: Status 404 returned error can't find the container with id b6ee9a80466069fa1e166317d123a0c4bf497b3097a1399069ca44a9effb32fd Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.992257 4838 generic.go:334] "Generic (PLEG): container finished" podID="298292bd-8186-4646-ad35-8058c9963946" containerID="2ba8fddc8a33dfaed82c785f863267073eb9608fab8ab17f0c11f872b5c6e6d8" exitCode=0 Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.993397 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ecf1-account-create-update-k27fv" event={"ID":"298292bd-8186-4646-ad35-8058c9963946","Type":"ContainerDied","Data":"2ba8fddc8a33dfaed82c785f863267073eb9608fab8ab17f0c11f872b5c6e6d8"} Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.995132 4838 generic.go:334] "Generic (PLEG): container finished" podID="6369855a-6de3-482c-8b9c-3e0263b7979f" containerID="3a81939f19a7888bbc7b1d965225c8d61c2b5dacb2513fc3c2531cf4b8c4cb07" exitCode=0 Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.995194 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5nft8" event={"ID":"6369855a-6de3-482c-8b9c-3e0263b7979f","Type":"ContainerDied","Data":"3a81939f19a7888bbc7b1d965225c8d61c2b5dacb2513fc3c2531cf4b8c4cb07"} Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.995216 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5nft8" event={"ID":"6369855a-6de3-482c-8b9c-3e0263b7979f","Type":"ContainerStarted","Data":"76c89137f95e6f4f7bc6cd7bb5e251d6a293de00a506398425168a957deef246"} Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.996455 4838 generic.go:334] "Generic (PLEG): container finished" podID="05852af0-5b52-4356-97e5-2fb3e9800c65" containerID="8082d8d1d591b7d583d27c21776e7508886c399bef9fdfed7413bfddb8a5cf47" exitCode=0 Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.996503 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-21a4-account-create-update-p44qz" event={"ID":"05852af0-5b52-4356-97e5-2fb3e9800c65","Type":"ContainerDied","Data":"8082d8d1d591b7d583d27c21776e7508886c399bef9fdfed7413bfddb8a5cf47"} Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.998028 4838 generic.go:334] "Generic (PLEG): container finished" podID="305e31d7-29e9-44f4-9dec-2aa2fd20c46c" containerID="30b95a7a6ec279e0cd2c3400099c92390b94d7f99da3170a4d4a204409f5d4bb" exitCode=0 Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.998077 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1db7-account-create-update-6m8pz" event={"ID":"305e31d7-29e9-44f4-9dec-2aa2fd20c46c","Type":"ContainerDied","Data":"30b95a7a6ec279e0cd2c3400099c92390b94d7f99da3170a4d4a204409f5d4bb"} Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.998220 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1db7-account-create-update-6m8pz" event={"ID":"305e31d7-29e9-44f4-9dec-2aa2fd20c46c","Type":"ContainerStarted","Data":"b6ee9a80466069fa1e166317d123a0c4bf497b3097a1399069ca44a9effb32fd"} Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.999315 4838 generic.go:334] "Generic (PLEG): container finished" podID="ee0e6d6c-351d-4b99-82d1-b2ad1042e44e" containerID="3dfe2c1926797d8cb64d8a414bc43c6b7980f204a0e84e8ebeab75650335e06b" exitCode=0 Dec 07 09:24:44 crc kubenswrapper[4838]: I1207 09:24:44.999381 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g9b52" event={"ID":"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e","Type":"ContainerDied","Data":"3dfe2c1926797d8cb64d8a414bc43c6b7980f204a0e84e8ebeab75650335e06b"} Dec 07 09:24:45 crc kubenswrapper[4838]: I1207 09:24:45.000492 4838 generic.go:334] "Generic (PLEG): container finished" podID="6863df38-7011-4203-adb9-117671f13bf4" containerID="2500aefe6979699207447925bb0966c448336da033795d2d1e03a87a69900cfb" exitCode=0 Dec 07 09:24:45 crc kubenswrapper[4838]: I1207 09:24:45.000612 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5jhzd" event={"ID":"6863df38-7011-4203-adb9-117671f13bf4","Type":"ContainerDied","Data":"2500aefe6979699207447925bb0966c448336da033795d2d1e03a87a69900cfb"} Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.439585 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5nft8" Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.543486 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6369855a-6de3-482c-8b9c-3e0263b7979f-operator-scripts\") pod \"6369855a-6de3-482c-8b9c-3e0263b7979f\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.543897 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffxk5\" (UniqueName: \"kubernetes.io/projected/6369855a-6de3-482c-8b9c-3e0263b7979f-kube-api-access-ffxk5\") pod \"6369855a-6de3-482c-8b9c-3e0263b7979f\" (UID: \"6369855a-6de3-482c-8b9c-3e0263b7979f\") " Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.546038 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6369855a-6de3-482c-8b9c-3e0263b7979f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6369855a-6de3-482c-8b9c-3e0263b7979f" (UID: "6369855a-6de3-482c-8b9c-3e0263b7979f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.558118 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6369855a-6de3-482c-8b9c-3e0263b7979f-kube-api-access-ffxk5" (OuterVolumeSpecName: "kube-api-access-ffxk5") pod "6369855a-6de3-482c-8b9c-3e0263b7979f" (UID: "6369855a-6de3-482c-8b9c-3e0263b7979f"). InnerVolumeSpecName "kube-api-access-ffxk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.581618 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.656181 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffxk5\" (UniqueName: \"kubernetes.io/projected/6369855a-6de3-482c-8b9c-3e0263b7979f-kube-api-access-ffxk5\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:46 crc kubenswrapper[4838]: I1207 09:24:46.656224 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6369855a-6de3-482c-8b9c-3e0263b7979f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:47 crc kubenswrapper[4838]: I1207 09:24:47.018239 4838 generic.go:334] "Generic (PLEG): container finished" podID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerID="73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2" exitCode=0 Dec 07 09:24:47 crc kubenswrapper[4838]: I1207 09:24:47.018313 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eae80b29-4c64-403b-be38-7fd30b7ba66f","Type":"ContainerDied","Data":"73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2"} Dec 07 09:24:47 crc kubenswrapper[4838]: I1207 09:24:47.020125 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5nft8" event={"ID":"6369855a-6de3-482c-8b9c-3e0263b7979f","Type":"ContainerDied","Data":"76c89137f95e6f4f7bc6cd7bb5e251d6a293de00a506398425168a957deef246"} Dec 07 09:24:47 crc kubenswrapper[4838]: I1207 09:24:47.020166 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76c89137f95e6f4f7bc6cd7bb5e251d6a293de00a506398425168a957deef246" Dec 07 09:24:47 crc kubenswrapper[4838]: I1207 09:24:47.020194 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5nft8" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.259205 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.274770 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.280222 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.285202 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-operator-scripts\") pod \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.285872 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82kbx\" (UniqueName: \"kubernetes.io/projected/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-kube-api-access-82kbx\") pod \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\" (UID: \"305e31d7-29e9-44f4-9dec-2aa2fd20c46c\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.286100 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "305e31d7-29e9-44f4-9dec-2aa2fd20c46c" (UID: "305e31d7-29e9-44f4-9dec-2aa2fd20c46c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.286549 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.291255 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-kube-api-access-82kbx" (OuterVolumeSpecName: "kube-api-access-82kbx") pod "305e31d7-29e9-44f4-9dec-2aa2fd20c46c" (UID: "305e31d7-29e9-44f4-9dec-2aa2fd20c46c"). InnerVolumeSpecName "kube-api-access-82kbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.345177 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.349714 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.387718 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-operator-scripts\") pod \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.387837 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05852af0-5b52-4356-97e5-2fb3e9800c65-operator-scripts\") pod \"05852af0-5b52-4356-97e5-2fb3e9800c65\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.387888 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6863df38-7011-4203-adb9-117671f13bf4-operator-scripts\") pod \"6863df38-7011-4203-adb9-117671f13bf4\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.387951 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpwdz\" (UniqueName: \"kubernetes.io/projected/298292bd-8186-4646-ad35-8058c9963946-kube-api-access-jpwdz\") pod \"298292bd-8186-4646-ad35-8058c9963946\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.387998 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd2vq\" (UniqueName: \"kubernetes.io/projected/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-kube-api-access-pd2vq\") pod \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\" (UID: \"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.388041 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8szk\" (UniqueName: \"kubernetes.io/projected/05852af0-5b52-4356-97e5-2fb3e9800c65-kube-api-access-z8szk\") pod \"05852af0-5b52-4356-97e5-2fb3e9800c65\" (UID: \"05852af0-5b52-4356-97e5-2fb3e9800c65\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.388086 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298292bd-8186-4646-ad35-8058c9963946-operator-scripts\") pod \"298292bd-8186-4646-ad35-8058c9963946\" (UID: \"298292bd-8186-4646-ad35-8058c9963946\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.388123 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msmhg\" (UniqueName: \"kubernetes.io/projected/6863df38-7011-4203-adb9-117671f13bf4-kube-api-access-msmhg\") pod \"6863df38-7011-4203-adb9-117671f13bf4\" (UID: \"6863df38-7011-4203-adb9-117671f13bf4\") " Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.388459 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82kbx\" (UniqueName: \"kubernetes.io/projected/305e31d7-29e9-44f4-9dec-2aa2fd20c46c-kube-api-access-82kbx\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.394369 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05852af0-5b52-4356-97e5-2fb3e9800c65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "05852af0-5b52-4356-97e5-2fb3e9800c65" (UID: "05852af0-5b52-4356-97e5-2fb3e9800c65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.394945 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee0e6d6c-351d-4b99-82d1-b2ad1042e44e" (UID: "ee0e6d6c-351d-4b99-82d1-b2ad1042e44e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.395545 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/298292bd-8186-4646-ad35-8058c9963946-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "298292bd-8186-4646-ad35-8058c9963946" (UID: "298292bd-8186-4646-ad35-8058c9963946"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.395920 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6863df38-7011-4203-adb9-117671f13bf4-kube-api-access-msmhg" (OuterVolumeSpecName: "kube-api-access-msmhg") pod "6863df38-7011-4203-adb9-117671f13bf4" (UID: "6863df38-7011-4203-adb9-117671f13bf4"). InnerVolumeSpecName "kube-api-access-msmhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.396107 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6863df38-7011-4203-adb9-117671f13bf4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6863df38-7011-4203-adb9-117671f13bf4" (UID: "6863df38-7011-4203-adb9-117671f13bf4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.397798 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/298292bd-8186-4646-ad35-8058c9963946-kube-api-access-jpwdz" (OuterVolumeSpecName: "kube-api-access-jpwdz") pod "298292bd-8186-4646-ad35-8058c9963946" (UID: "298292bd-8186-4646-ad35-8058c9963946"). InnerVolumeSpecName "kube-api-access-jpwdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.403711 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-kube-api-access-pd2vq" (OuterVolumeSpecName: "kube-api-access-pd2vq") pod "ee0e6d6c-351d-4b99-82d1-b2ad1042e44e" (UID: "ee0e6d6c-351d-4b99-82d1-b2ad1042e44e"). InnerVolumeSpecName "kube-api-access-pd2vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.404964 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05852af0-5b52-4356-97e5-2fb3e9800c65-kube-api-access-z8szk" (OuterVolumeSpecName: "kube-api-access-z8szk") pod "05852af0-5b52-4356-97e5-2fb3e9800c65" (UID: "05852af0-5b52-4356-97e5-2fb3e9800c65"). InnerVolumeSpecName "kube-api-access-z8szk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489493 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298292bd-8186-4646-ad35-8058c9963946-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489538 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msmhg\" (UniqueName: \"kubernetes.io/projected/6863df38-7011-4203-adb9-117671f13bf4-kube-api-access-msmhg\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489552 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489563 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05852af0-5b52-4356-97e5-2fb3e9800c65-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489575 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6863df38-7011-4203-adb9-117671f13bf4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489588 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpwdz\" (UniqueName: \"kubernetes.io/projected/298292bd-8186-4646-ad35-8058c9963946-kube-api-access-jpwdz\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489599 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd2vq\" (UniqueName: \"kubernetes.io/projected/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e-kube-api-access-pd2vq\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:48 crc kubenswrapper[4838]: I1207 09:24:48.489608 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8szk\" (UniqueName: \"kubernetes.io/projected/05852af0-5b52-4356-97e5-2fb3e9800c65-kube-api-access-z8szk\") on node \"crc\" DevicePath \"\"" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.040311 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-21a4-account-create-update-p44qz" event={"ID":"05852af0-5b52-4356-97e5-2fb3e9800c65","Type":"ContainerDied","Data":"3e81fc388eb254e8fb282ba5ab3b5907ae05a3e198e8abe7bb1ff193660d35da"} Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.040593 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e81fc388eb254e8fb282ba5ab3b5907ae05a3e198e8abe7bb1ff193660d35da" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.040321 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-21a4-account-create-update-p44qz" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.041781 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1db7-account-create-update-6m8pz" event={"ID":"305e31d7-29e9-44f4-9dec-2aa2fd20c46c","Type":"ContainerDied","Data":"b6ee9a80466069fa1e166317d123a0c4bf497b3097a1399069ca44a9effb32fd"} Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.041888 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6ee9a80466069fa1e166317d123a0c4bf497b3097a1399069ca44a9effb32fd" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.041805 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1db7-account-create-update-6m8pz" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.043843 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g9b52" event={"ID":"ee0e6d6c-351d-4b99-82d1-b2ad1042e44e","Type":"ContainerDied","Data":"011ebbfca5ecd24e78fdfa529461635975c9ef1f032756a50e1ab7965ecd7303"} Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.043922 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="011ebbfca5ecd24e78fdfa529461635975c9ef1f032756a50e1ab7965ecd7303" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.043891 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9b52" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.048299 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5jhzd" event={"ID":"6863df38-7011-4203-adb9-117671f13bf4","Type":"ContainerDied","Data":"01cfc8331c9c0f91485029bce4984a84afa9946b538ef3fa44287f78ba0fe899"} Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.048347 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01cfc8331c9c0f91485029bce4984a84afa9946b538ef3fa44287f78ba0fe899" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.048309 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jhzd" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.049947 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ecf1-account-create-update-k27fv" event={"ID":"298292bd-8186-4646-ad35-8058c9963946","Type":"ContainerDied","Data":"7d1f5716edabe5d7d8c7fbda787917078077c31d03aa40086816c2272e11062d"} Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.049990 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d1f5716edabe5d7d8c7fbda787917078077c31d03aa40086816c2272e11062d" Dec 07 09:24:49 crc kubenswrapper[4838]: I1207 09:24:49.050137 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ecf1-account-create-update-k27fv" Dec 07 09:24:50 crc kubenswrapper[4838]: I1207 09:24:50.058757 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eae80b29-4c64-403b-be38-7fd30b7ba66f","Type":"ContainerStarted","Data":"269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004"} Dec 07 09:24:50 crc kubenswrapper[4838]: I1207 09:24:50.059298 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 07 09:24:50 crc kubenswrapper[4838]: I1207 09:24:50.087658 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.240208312 podStartE2EDuration="1m12.087637339s" podCreationTimestamp="2025-12-07 09:23:38 +0000 UTC" firstStartedPulling="2025-12-07 09:23:40.401872774 +0000 UTC m=+1037.109191791" lastFinishedPulling="2025-12-07 09:24:13.249301801 +0000 UTC m=+1069.956620818" observedRunningTime="2025-12-07 09:24:50.083114871 +0000 UTC m=+1106.790433888" watchObservedRunningTime="2025-12-07 09:24:50.087637339 +0000 UTC m=+1106.794956356" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.404557 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-n2jzr"] Dec 07 09:24:53 crc kubenswrapper[4838]: E1207 09:24:53.405196 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298292bd-8186-4646-ad35-8058c9963946" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405212 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="298292bd-8186-4646-ad35-8058c9963946" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: E1207 09:24:53.405233 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6863df38-7011-4203-adb9-117671f13bf4" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405240 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6863df38-7011-4203-adb9-117671f13bf4" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: E1207 09:24:53.405266 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05852af0-5b52-4356-97e5-2fb3e9800c65" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405275 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="05852af0-5b52-4356-97e5-2fb3e9800c65" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: E1207 09:24:53.405293 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6369855a-6de3-482c-8b9c-3e0263b7979f" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405300 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6369855a-6de3-482c-8b9c-3e0263b7979f" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: E1207 09:24:53.405311 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305e31d7-29e9-44f4-9dec-2aa2fd20c46c" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405319 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="305e31d7-29e9-44f4-9dec-2aa2fd20c46c" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: E1207 09:24:53.405333 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0e6d6c-351d-4b99-82d1-b2ad1042e44e" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405342 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0e6d6c-351d-4b99-82d1-b2ad1042e44e" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405508 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="298292bd-8186-4646-ad35-8058c9963946" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405525 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee0e6d6c-351d-4b99-82d1-b2ad1042e44e" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405536 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="05852af0-5b52-4356-97e5-2fb3e9800c65" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405548 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6369855a-6de3-482c-8b9c-3e0263b7979f" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405560 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="305e31d7-29e9-44f4-9dec-2aa2fd20c46c" containerName="mariadb-account-create-update" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.405569 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6863df38-7011-4203-adb9-117671f13bf4" containerName="mariadb-database-create" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.406222 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.408027 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pw4f6" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.408406 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.411208 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-n2jzr"] Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.505084 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-sbl7t" podUID="a3ae58f2-6cd2-4782-96b1-a88597b25b37" containerName="ovn-controller" probeResult="failure" output=< Dec 07 09:24:53 crc kubenswrapper[4838]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 07 09:24:53 crc kubenswrapper[4838]: > Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.520385 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-combined-ca-bundle\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.520474 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-config-data\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.520559 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-db-sync-config-data\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.520597 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgrgv\" (UniqueName: \"kubernetes.io/projected/eb3152f7-2ea4-4de8-ba34-038fd733682b-kube-api-access-mgrgv\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.621623 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-config-data\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.621974 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-db-sync-config-data\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.622105 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgrgv\" (UniqueName: \"kubernetes.io/projected/eb3152f7-2ea4-4de8-ba34-038fd733682b-kube-api-access-mgrgv\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.622270 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-combined-ca-bundle\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.627788 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-combined-ca-bundle\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.628450 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-db-sync-config-data\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.629436 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-config-data\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.639383 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgrgv\" (UniqueName: \"kubernetes.io/projected/eb3152f7-2ea4-4de8-ba34-038fd733682b-kube-api-access-mgrgv\") pod \"glance-db-sync-n2jzr\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:53 crc kubenswrapper[4838]: I1207 09:24:53.727203 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n2jzr" Dec 07 09:24:54 crc kubenswrapper[4838]: I1207 09:24:54.333332 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-n2jzr"] Dec 07 09:24:54 crc kubenswrapper[4838]: I1207 09:24:54.492735 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:24:54 crc kubenswrapper[4838]: I1207 09:24:54.492856 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:24:55 crc kubenswrapper[4838]: I1207 09:24:55.098430 4838 generic.go:334] "Generic (PLEG): container finished" podID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerID="fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7" exitCode=0 Dec 07 09:24:55 crc kubenswrapper[4838]: I1207 09:24:55.098731 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454","Type":"ContainerDied","Data":"fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7"} Dec 07 09:24:55 crc kubenswrapper[4838]: I1207 09:24:55.101025 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n2jzr" event={"ID":"eb3152f7-2ea4-4de8-ba34-038fd733682b","Type":"ContainerStarted","Data":"5d3dc0426a95197590360bff6fe52fe2f719b28428514bf736be03ba508fa206"} Dec 07 09:24:56 crc kubenswrapper[4838]: I1207 09:24:56.110132 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454","Type":"ContainerStarted","Data":"a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03"} Dec 07 09:24:56 crc kubenswrapper[4838]: I1207 09:24:56.110574 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:24:56 crc kubenswrapper[4838]: I1207 09:24:56.136939 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371958.717852 podStartE2EDuration="1m18.136924872s" podCreationTimestamp="2025-12-07 09:23:38 +0000 UTC" firstStartedPulling="2025-12-07 09:23:40.027584166 +0000 UTC m=+1036.734903183" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:24:56.133921097 +0000 UTC m=+1112.841240114" watchObservedRunningTime="2025-12-07 09:24:56.136924872 +0000 UTC m=+1112.844243889" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.533350 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-sbl7t" podUID="a3ae58f2-6cd2-4782-96b1-a88597b25b37" containerName="ovn-controller" probeResult="failure" output=< Dec 07 09:24:58 crc kubenswrapper[4838]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 07 09:24:58 crc kubenswrapper[4838]: > Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.555970 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.565243 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gb65g" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.834747 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sbl7t-config-n4sbw"] Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.836059 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.838706 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.850411 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbl7t-config-n4sbw"] Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.907772 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-additional-scripts\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.908215 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-log-ovn\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.908260 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwt5t\" (UniqueName: \"kubernetes.io/projected/db988bef-e07b-448a-92c3-1c126452763e-kube-api-access-jwt5t\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.908346 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.908394 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-scripts\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:58 crc kubenswrapper[4838]: I1207 09:24:58.908462 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run-ovn\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.010560 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.010619 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-scripts\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.010661 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run-ovn\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.010745 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-additional-scripts\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.010793 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-log-ovn\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.010847 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwt5t\" (UniqueName: \"kubernetes.io/projected/db988bef-e07b-448a-92c3-1c126452763e-kube-api-access-jwt5t\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.011316 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.011784 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-log-ovn\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.011973 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-additional-scripts\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.013281 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-scripts\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.013372 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run-ovn\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.046960 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwt5t\" (UniqueName: \"kubernetes.io/projected/db988bef-e07b-448a-92c3-1c126452763e-kube-api-access-jwt5t\") pod \"ovn-controller-sbl7t-config-n4sbw\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.166311 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.692639 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbl7t-config-n4sbw"] Dec 07 09:24:59 crc kubenswrapper[4838]: I1207 09:24:59.894057 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.346386 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xql5b"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.347407 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.373918 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-f574-account-create-update-8q9qx"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.375015 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.378693 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.384305 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xql5b"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.398998 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f574-account-create-update-8q9qx"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.434288 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42ljn\" (UniqueName: \"kubernetes.io/projected/b524180a-e6ae-4cb3-9717-99ad184944b8-kube-api-access-42ljn\") pod \"cinder-db-create-xql5b\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.434348 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55tnb\" (UniqueName: \"kubernetes.io/projected/c2a91f97-d03b-4da9-b70c-e858e878470f-kube-api-access-55tnb\") pod \"barbican-f574-account-create-update-8q9qx\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.434411 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2a91f97-d03b-4da9-b70c-e858e878470f-operator-scripts\") pod \"barbican-f574-account-create-update-8q9qx\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.434426 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b524180a-e6ae-4cb3-9717-99ad184944b8-operator-scripts\") pod \"cinder-db-create-xql5b\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.466360 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-cxm29"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.468050 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.496413 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cxm29"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.536682 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55tnb\" (UniqueName: \"kubernetes.io/projected/c2a91f97-d03b-4da9-b70c-e858e878470f-kube-api-access-55tnb\") pod \"barbican-f574-account-create-update-8q9qx\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.536755 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-operator-scripts\") pod \"barbican-db-create-cxm29\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.536791 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flccn\" (UniqueName: \"kubernetes.io/projected/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-kube-api-access-flccn\") pod \"barbican-db-create-cxm29\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.536837 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2a91f97-d03b-4da9-b70c-e858e878470f-operator-scripts\") pod \"barbican-f574-account-create-update-8q9qx\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.536859 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b524180a-e6ae-4cb3-9717-99ad184944b8-operator-scripts\") pod \"cinder-db-create-xql5b\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.536926 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42ljn\" (UniqueName: \"kubernetes.io/projected/b524180a-e6ae-4cb3-9717-99ad184944b8-kube-api-access-42ljn\") pod \"cinder-db-create-xql5b\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.537994 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2a91f97-d03b-4da9-b70c-e858e878470f-operator-scripts\") pod \"barbican-f574-account-create-update-8q9qx\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.538018 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b524180a-e6ae-4cb3-9717-99ad184944b8-operator-scripts\") pod \"cinder-db-create-xql5b\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.564715 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55tnb\" (UniqueName: \"kubernetes.io/projected/c2a91f97-d03b-4da9-b70c-e858e878470f-kube-api-access-55tnb\") pod \"barbican-f574-account-create-update-8q9qx\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.569644 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42ljn\" (UniqueName: \"kubernetes.io/projected/b524180a-e6ae-4cb3-9717-99ad184944b8-kube-api-access-42ljn\") pod \"cinder-db-create-xql5b\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.584872 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-72ed-account-create-update-tx6p4"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.585752 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.596099 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.638322 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-operator-scripts\") pod \"barbican-db-create-cxm29\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.638373 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flccn\" (UniqueName: \"kubernetes.io/projected/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-kube-api-access-flccn\") pod \"barbican-db-create-cxm29\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.638417 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsvlt\" (UniqueName: \"kubernetes.io/projected/a38f9117-4655-46af-8b54-01ca6b4e6553-kube-api-access-xsvlt\") pod \"cinder-72ed-account-create-update-tx6p4\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.638440 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f9117-4655-46af-8b54-01ca6b4e6553-operator-scripts\") pod \"cinder-72ed-account-create-update-tx6p4\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.639359 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-operator-scripts\") pod \"barbican-db-create-cxm29\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.660357 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-kq6lf"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.667059 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.668987 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.669710 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-72ed-account-create-update-tx6p4"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.685225 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-kq6lf"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.688062 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.689532 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flccn\" (UniqueName: \"kubernetes.io/projected/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-kube-api-access-flccn\") pod \"barbican-db-create-cxm29\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.739530 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsvlt\" (UniqueName: \"kubernetes.io/projected/a38f9117-4655-46af-8b54-01ca6b4e6553-kube-api-access-xsvlt\") pod \"cinder-72ed-account-create-update-tx6p4\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.739944 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6h4m\" (UniqueName: \"kubernetes.io/projected/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-kube-api-access-v6h4m\") pod \"neutron-db-create-kq6lf\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.739990 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f9117-4655-46af-8b54-01ca6b4e6553-operator-scripts\") pod \"cinder-72ed-account-create-update-tx6p4\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.740210 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-operator-scripts\") pod \"neutron-db-create-kq6lf\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.741056 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f9117-4655-46af-8b54-01ca6b4e6553-operator-scripts\") pod \"cinder-72ed-account-create-update-tx6p4\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.785925 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.792445 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-rnxlk"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.793604 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.798247 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v7gbr" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.798512 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.798986 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.804189 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsvlt\" (UniqueName: \"kubernetes.io/projected/a38f9117-4655-46af-8b54-01ca6b4e6553-kube-api-access-xsvlt\") pod \"cinder-72ed-account-create-update-tx6p4\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.818237 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rnxlk"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.818690 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.850803 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-operator-scripts\") pod \"neutron-db-create-kq6lf\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.850863 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6h4m\" (UniqueName: \"kubernetes.io/projected/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-kube-api-access-v6h4m\") pod \"neutron-db-create-kq6lf\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.850897 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-combined-ca-bundle\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.850965 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnqqk\" (UniqueName: \"kubernetes.io/projected/aa78bff1-c04d-42de-a3f1-7019500bd194-kube-api-access-wnqqk\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.851019 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-config-data\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.851626 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-operator-scripts\") pod \"neutron-db-create-kq6lf\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.877538 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-800d-account-create-update-5qbg4"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.879463 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.881250 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.897432 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6h4m\" (UniqueName: \"kubernetes.io/projected/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-kube-api-access-v6h4m\") pod \"neutron-db-create-kq6lf\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.907477 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-800d-account-create-update-5qbg4"] Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.949192 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.953019 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnqqk\" (UniqueName: \"kubernetes.io/projected/aa78bff1-c04d-42de-a3f1-7019500bd194-kube-api-access-wnqqk\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.953133 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgxrg\" (UniqueName: \"kubernetes.io/projected/344ac587-0d87-4812-9d65-18fc79eec28b-kube-api-access-kgxrg\") pod \"neutron-800d-account-create-update-5qbg4\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.953178 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-config-data\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.953248 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-combined-ca-bundle\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.953285 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344ac587-0d87-4812-9d65-18fc79eec28b-operator-scripts\") pod \"neutron-800d-account-create-update-5qbg4\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.958319 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-combined-ca-bundle\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.960220 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-config-data\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:00 crc kubenswrapper[4838]: I1207 09:25:00.973119 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnqqk\" (UniqueName: \"kubernetes.io/projected/aa78bff1-c04d-42de-a3f1-7019500bd194-kube-api-access-wnqqk\") pod \"keystone-db-sync-rnxlk\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:01 crc kubenswrapper[4838]: I1207 09:25:01.026702 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:01 crc kubenswrapper[4838]: I1207 09:25:01.054696 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344ac587-0d87-4812-9d65-18fc79eec28b-operator-scripts\") pod \"neutron-800d-account-create-update-5qbg4\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:01 crc kubenswrapper[4838]: I1207 09:25:01.054824 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgxrg\" (UniqueName: \"kubernetes.io/projected/344ac587-0d87-4812-9d65-18fc79eec28b-kube-api-access-kgxrg\") pod \"neutron-800d-account-create-update-5qbg4\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:01 crc kubenswrapper[4838]: I1207 09:25:01.055803 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344ac587-0d87-4812-9d65-18fc79eec28b-operator-scripts\") pod \"neutron-800d-account-create-update-5qbg4\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:01 crc kubenswrapper[4838]: I1207 09:25:01.073100 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgxrg\" (UniqueName: \"kubernetes.io/projected/344ac587-0d87-4812-9d65-18fc79eec28b-kube-api-access-kgxrg\") pod \"neutron-800d-account-create-update-5qbg4\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:01 crc kubenswrapper[4838]: I1207 09:25:01.140675 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:01 crc kubenswrapper[4838]: I1207 09:25:01.238921 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:03 crc kubenswrapper[4838]: I1207 09:25:03.508686 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-sbl7t" podUID="a3ae58f2-6cd2-4782-96b1-a88597b25b37" containerName="ovn-controller" probeResult="failure" output=< Dec 07 09:25:03 crc kubenswrapper[4838]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 07 09:25:03 crc kubenswrapper[4838]: > Dec 07 09:25:07 crc kubenswrapper[4838]: I1207 09:25:07.888727 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xql5b"] Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.010690 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-800d-account-create-update-5qbg4"] Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.018922 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cxm29"] Dec 07 09:25:08 crc kubenswrapper[4838]: W1207 09:25:08.023273 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddcb1b9bd_baf9_44fa_b43a_a181ffcaf89c.slice/crio-ac1b6e86fff3aa0d07d20ecf44305e56a11ecc752530a133564d2227228a60e1 WatchSource:0}: Error finding container ac1b6e86fff3aa0d07d20ecf44305e56a11ecc752530a133564d2227228a60e1: Status 404 returned error can't find the container with id ac1b6e86fff3aa0d07d20ecf44305e56a11ecc752530a133564d2227228a60e1 Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.124856 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-72ed-account-create-update-tx6p4"] Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.153237 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rnxlk"] Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.170910 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f574-account-create-update-8q9qx"] Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.176169 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-kq6lf"] Dec 07 09:25:08 crc kubenswrapper[4838]: W1207 09:25:08.206204 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2a91f97_d03b_4da9_b70c_e858e878470f.slice/crio-2bd10724c8dd38e5549d8e57576159794f5f1b9e1d0c6988ce612015bced9ddf WatchSource:0}: Error finding container 2bd10724c8dd38e5549d8e57576159794f5f1b9e1d0c6988ce612015bced9ddf: Status 404 returned error can't find the container with id 2bd10724c8dd38e5549d8e57576159794f5f1b9e1d0c6988ce612015bced9ddf Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.206230 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kq6lf" event={"ID":"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f","Type":"ContainerStarted","Data":"1aa2760a8513ec844ebd60e8a8f77df8ff235806016e44b8143c10413aec00b2"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.208143 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72ed-account-create-update-tx6p4" event={"ID":"a38f9117-4655-46af-8b54-01ca6b4e6553","Type":"ContainerStarted","Data":"513332bd125ae9deae567dfa1b7c1a76a103366300de4751eff2c705f1b31635"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.209507 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t-config-n4sbw" event={"ID":"db988bef-e07b-448a-92c3-1c126452763e","Type":"ContainerStarted","Data":"919e11cc574bf359d79f319331a94db2ff6666f563e4ce270263c73f88d7f826"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.209540 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t-config-n4sbw" event={"ID":"db988bef-e07b-448a-92c3-1c126452763e","Type":"ContainerStarted","Data":"7781122133ce64b614272d3d0ac1276d39f2256aadb8ce0d4624548359adfc61"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.211106 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-800d-account-create-update-5qbg4" event={"ID":"344ac587-0d87-4812-9d65-18fc79eec28b","Type":"ContainerStarted","Data":"43f8f10b48da146403ecf8c711a3ae1eff7b624c6ff059a376919c43883899d3"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.220279 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cxm29" event={"ID":"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c","Type":"ContainerStarted","Data":"ac1b6e86fff3aa0d07d20ecf44305e56a11ecc752530a133564d2227228a60e1"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.222612 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnxlk" event={"ID":"aa78bff1-c04d-42de-a3f1-7019500bd194","Type":"ContainerStarted","Data":"09d2c5860ef52a06064468a846c0dab7ebbfa71faa826f5204a9003c040b971b"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.233897 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-sbl7t-config-n4sbw" podStartSLOduration=10.233882782 podStartE2EDuration="10.233882782s" podCreationTimestamp="2025-12-07 09:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:08.228768218 +0000 UTC m=+1124.936087235" watchObservedRunningTime="2025-12-07 09:25:08.233882782 +0000 UTC m=+1124.941201789" Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.237031 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xql5b" event={"ID":"b524180a-e6ae-4cb3-9717-99ad184944b8","Type":"ContainerStarted","Data":"f014559b089827c91f803ee38dd392f758e3143e9253a23d798a766e65f58e3f"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.237071 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xql5b" event={"ID":"b524180a-e6ae-4cb3-9717-99ad184944b8","Type":"ContainerStarted","Data":"f3dd70b072c20173851308a30d9f50c45502c3b8e51c26815cb8cdee353bafb9"} Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.262221 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-xql5b" podStartSLOduration=8.262199263 podStartE2EDuration="8.262199263s" podCreationTimestamp="2025-12-07 09:25:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:08.253286641 +0000 UTC m=+1124.960605668" watchObservedRunningTime="2025-12-07 09:25:08.262199263 +0000 UTC m=+1124.969518280" Dec 07 09:25:08 crc kubenswrapper[4838]: I1207 09:25:08.549325 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-sbl7t" Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.247794 4838 generic.go:334] "Generic (PLEG): container finished" podID="b524180a-e6ae-4cb3-9717-99ad184944b8" containerID="f014559b089827c91f803ee38dd392f758e3143e9253a23d798a766e65f58e3f" exitCode=0 Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.247924 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xql5b" event={"ID":"b524180a-e6ae-4cb3-9717-99ad184944b8","Type":"ContainerDied","Data":"f014559b089827c91f803ee38dd392f758e3143e9253a23d798a766e65f58e3f"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.255849 4838 generic.go:334] "Generic (PLEG): container finished" podID="c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f" containerID="a1063ec138e1b08dc2194a29b7f6adf93140bb578db8fea968b0c3cb3aadce30" exitCode=0 Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.255942 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kq6lf" event={"ID":"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f","Type":"ContainerDied","Data":"a1063ec138e1b08dc2194a29b7f6adf93140bb578db8fea968b0c3cb3aadce30"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.258834 4838 generic.go:334] "Generic (PLEG): container finished" podID="a38f9117-4655-46af-8b54-01ca6b4e6553" containerID="60b2ac53a6c707e42f86895b236f2b648d2910e9a5365f1f0e18338f94c669c2" exitCode=0 Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.258878 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72ed-account-create-update-tx6p4" event={"ID":"a38f9117-4655-46af-8b54-01ca6b4e6553","Type":"ContainerDied","Data":"60b2ac53a6c707e42f86895b236f2b648d2910e9a5365f1f0e18338f94c669c2"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.268957 4838 generic.go:334] "Generic (PLEG): container finished" podID="c2a91f97-d03b-4da9-b70c-e858e878470f" containerID="513f58981ccaa9541916bf7c609645fb39ce6baf276662d9ca7a98f31aac9bef" exitCode=0 Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.269015 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f574-account-create-update-8q9qx" event={"ID":"c2a91f97-d03b-4da9-b70c-e858e878470f","Type":"ContainerDied","Data":"513f58981ccaa9541916bf7c609645fb39ce6baf276662d9ca7a98f31aac9bef"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.269041 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f574-account-create-update-8q9qx" event={"ID":"c2a91f97-d03b-4da9-b70c-e858e878470f","Type":"ContainerStarted","Data":"2bd10724c8dd38e5549d8e57576159794f5f1b9e1d0c6988ce612015bced9ddf"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.272229 4838 generic.go:334] "Generic (PLEG): container finished" podID="db988bef-e07b-448a-92c3-1c126452763e" containerID="919e11cc574bf359d79f319331a94db2ff6666f563e4ce270263c73f88d7f826" exitCode=0 Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.272369 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t-config-n4sbw" event={"ID":"db988bef-e07b-448a-92c3-1c126452763e","Type":"ContainerDied","Data":"919e11cc574bf359d79f319331a94db2ff6666f563e4ce270263c73f88d7f826"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.274365 4838 generic.go:334] "Generic (PLEG): container finished" podID="344ac587-0d87-4812-9d65-18fc79eec28b" containerID="d4704ab1072a5d251238bb238e97298c48190866c5e02967d1cbe4f202871dd0" exitCode=0 Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.274467 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-800d-account-create-update-5qbg4" event={"ID":"344ac587-0d87-4812-9d65-18fc79eec28b","Type":"ContainerDied","Data":"d4704ab1072a5d251238bb238e97298c48190866c5e02967d1cbe4f202871dd0"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.280464 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n2jzr" event={"ID":"eb3152f7-2ea4-4de8-ba34-038fd733682b","Type":"ContainerStarted","Data":"366f9728572c990684bf4b94aca2d5853c6b60bf195c3f16de285ed8133c3a41"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.282867 4838 generic.go:334] "Generic (PLEG): container finished" podID="dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c" containerID="77b7638595e61d5be645e33e577c052000890a26a36f4590d525ac5e338be56e" exitCode=0 Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.282894 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cxm29" event={"ID":"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c","Type":"ContainerDied","Data":"77b7638595e61d5be645e33e577c052000890a26a36f4590d525ac5e338be56e"} Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.380905 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-n2jzr" podStartSLOduration=3.2719899310000002 podStartE2EDuration="16.380883171s" podCreationTimestamp="2025-12-07 09:24:53 +0000 UTC" firstStartedPulling="2025-12-07 09:24:54.337083065 +0000 UTC m=+1111.044402082" lastFinishedPulling="2025-12-07 09:25:07.445976305 +0000 UTC m=+1124.153295322" observedRunningTime="2025-12-07 09:25:09.376125297 +0000 UTC m=+1126.083444314" watchObservedRunningTime="2025-12-07 09:25:09.380883171 +0000 UTC m=+1126.088202188" Dec 07 09:25:09 crc kubenswrapper[4838]: I1207 09:25:09.428073 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.726399 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.840591 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b524180a-e6ae-4cb3-9717-99ad184944b8-operator-scripts\") pod \"b524180a-e6ae-4cb3-9717-99ad184944b8\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.840711 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42ljn\" (UniqueName: \"kubernetes.io/projected/b524180a-e6ae-4cb3-9717-99ad184944b8-kube-api-access-42ljn\") pod \"b524180a-e6ae-4cb3-9717-99ad184944b8\" (UID: \"b524180a-e6ae-4cb3-9717-99ad184944b8\") " Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.842460 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b524180a-e6ae-4cb3-9717-99ad184944b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b524180a-e6ae-4cb3-9717-99ad184944b8" (UID: "b524180a-e6ae-4cb3-9717-99ad184944b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.852639 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b524180a-e6ae-4cb3-9717-99ad184944b8-kube-api-access-42ljn" (OuterVolumeSpecName: "kube-api-access-42ljn") pod "b524180a-e6ae-4cb3-9717-99ad184944b8" (UID: "b524180a-e6ae-4cb3-9717-99ad184944b8"). InnerVolumeSpecName "kube-api-access-42ljn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.937134 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.942345 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b524180a-e6ae-4cb3-9717-99ad184944b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.942371 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42ljn\" (UniqueName: \"kubernetes.io/projected/b524180a-e6ae-4cb3-9717-99ad184944b8-kube-api-access-42ljn\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.945439 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.965891 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:10 crc kubenswrapper[4838]: I1207 09:25:10.979437 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:10.996795 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049270 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2a91f97-d03b-4da9-b70c-e858e878470f-operator-scripts\") pod \"c2a91f97-d03b-4da9-b70c-e858e878470f\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049330 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flccn\" (UniqueName: \"kubernetes.io/projected/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-kube-api-access-flccn\") pod \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049401 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344ac587-0d87-4812-9d65-18fc79eec28b-operator-scripts\") pod \"344ac587-0d87-4812-9d65-18fc79eec28b\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049471 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55tnb\" (UniqueName: \"kubernetes.io/projected/c2a91f97-d03b-4da9-b70c-e858e878470f-kube-api-access-55tnb\") pod \"c2a91f97-d03b-4da9-b70c-e858e878470f\" (UID: \"c2a91f97-d03b-4da9-b70c-e858e878470f\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049493 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgxrg\" (UniqueName: \"kubernetes.io/projected/344ac587-0d87-4812-9d65-18fc79eec28b-kube-api-access-kgxrg\") pod \"344ac587-0d87-4812-9d65-18fc79eec28b\" (UID: \"344ac587-0d87-4812-9d65-18fc79eec28b\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049543 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-operator-scripts\") pod \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049586 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f9117-4655-46af-8b54-01ca6b4e6553-operator-scripts\") pod \"a38f9117-4655-46af-8b54-01ca6b4e6553\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049620 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-operator-scripts\") pod \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\" (UID: \"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049649 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsvlt\" (UniqueName: \"kubernetes.io/projected/a38f9117-4655-46af-8b54-01ca6b4e6553-kube-api-access-xsvlt\") pod \"a38f9117-4655-46af-8b54-01ca6b4e6553\" (UID: \"a38f9117-4655-46af-8b54-01ca6b4e6553\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.049663 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6h4m\" (UniqueName: \"kubernetes.io/projected/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-kube-api-access-v6h4m\") pod \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\" (UID: \"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.053298 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f" (UID: "c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.053693 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a38f9117-4655-46af-8b54-01ca6b4e6553-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a38f9117-4655-46af-8b54-01ca6b4e6553" (UID: "a38f9117-4655-46af-8b54-01ca6b4e6553"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.056403 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/344ac587-0d87-4812-9d65-18fc79eec28b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "344ac587-0d87-4812-9d65-18fc79eec28b" (UID: "344ac587-0d87-4812-9d65-18fc79eec28b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.054617 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2a91f97-d03b-4da9-b70c-e858e878470f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2a91f97-d03b-4da9-b70c-e858e878470f" (UID: "c2a91f97-d03b-4da9-b70c-e858e878470f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.056029 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-kube-api-access-v6h4m" (OuterVolumeSpecName: "kube-api-access-v6h4m") pod "c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f" (UID: "c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f"). InnerVolumeSpecName "kube-api-access-v6h4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.056396 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c" (UID: "dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.057360 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/344ac587-0d87-4812-9d65-18fc79eec28b-kube-api-access-kgxrg" (OuterVolumeSpecName: "kube-api-access-kgxrg") pod "344ac587-0d87-4812-9d65-18fc79eec28b" (UID: "344ac587-0d87-4812-9d65-18fc79eec28b"). InnerVolumeSpecName "kube-api-access-kgxrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.058343 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-kube-api-access-flccn" (OuterVolumeSpecName: "kube-api-access-flccn") pod "dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c" (UID: "dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c"). InnerVolumeSpecName "kube-api-access-flccn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.058681 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a91f97-d03b-4da9-b70c-e858e878470f-kube-api-access-55tnb" (OuterVolumeSpecName: "kube-api-access-55tnb") pod "c2a91f97-d03b-4da9-b70c-e858e878470f" (UID: "c2a91f97-d03b-4da9-b70c-e858e878470f"). InnerVolumeSpecName "kube-api-access-55tnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.072429 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38f9117-4655-46af-8b54-01ca6b4e6553-kube-api-access-xsvlt" (OuterVolumeSpecName: "kube-api-access-xsvlt") pod "a38f9117-4655-46af-8b54-01ca6b4e6553" (UID: "a38f9117-4655-46af-8b54-01ca6b4e6553"). InnerVolumeSpecName "kube-api-access-xsvlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.074322 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.150754 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwt5t\" (UniqueName: \"kubernetes.io/projected/db988bef-e07b-448a-92c3-1c126452763e-kube-api-access-jwt5t\") pod \"db988bef-e07b-448a-92c3-1c126452763e\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.150805 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-log-ovn\") pod \"db988bef-e07b-448a-92c3-1c126452763e\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.150865 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run-ovn\") pod \"db988bef-e07b-448a-92c3-1c126452763e\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.150887 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-additional-scripts\") pod \"db988bef-e07b-448a-92c3-1c126452763e\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151043 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run\") pod \"db988bef-e07b-448a-92c3-1c126452763e\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151078 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-scripts\") pod \"db988bef-e07b-448a-92c3-1c126452763e\" (UID: \"db988bef-e07b-448a-92c3-1c126452763e\") " Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151448 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344ac587-0d87-4812-9d65-18fc79eec28b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151467 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55tnb\" (UniqueName: \"kubernetes.io/projected/c2a91f97-d03b-4da9-b70c-e858e878470f-kube-api-access-55tnb\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151479 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgxrg\" (UniqueName: \"kubernetes.io/projected/344ac587-0d87-4812-9d65-18fc79eec28b-kube-api-access-kgxrg\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151489 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151500 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f9117-4655-46af-8b54-01ca6b4e6553-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151510 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151520 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsvlt\" (UniqueName: \"kubernetes.io/projected/a38f9117-4655-46af-8b54-01ca6b4e6553-kube-api-access-xsvlt\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151530 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6h4m\" (UniqueName: \"kubernetes.io/projected/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f-kube-api-access-v6h4m\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151540 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2a91f97-d03b-4da9-b70c-e858e878470f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.151551 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flccn\" (UniqueName: \"kubernetes.io/projected/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c-kube-api-access-flccn\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.152937 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-scripts" (OuterVolumeSpecName: "scripts") pod "db988bef-e07b-448a-92c3-1c126452763e" (UID: "db988bef-e07b-448a-92c3-1c126452763e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.153302 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "db988bef-e07b-448a-92c3-1c126452763e" (UID: "db988bef-e07b-448a-92c3-1c126452763e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.153344 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "db988bef-e07b-448a-92c3-1c126452763e" (UID: "db988bef-e07b-448a-92c3-1c126452763e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.153704 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "db988bef-e07b-448a-92c3-1c126452763e" (UID: "db988bef-e07b-448a-92c3-1c126452763e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.153742 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run" (OuterVolumeSpecName: "var-run") pod "db988bef-e07b-448a-92c3-1c126452763e" (UID: "db988bef-e07b-448a-92c3-1c126452763e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.156258 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db988bef-e07b-448a-92c3-1c126452763e-kube-api-access-jwt5t" (OuterVolumeSpecName: "kube-api-access-jwt5t") pod "db988bef-e07b-448a-92c3-1c126452763e" (UID: "db988bef-e07b-448a-92c3-1c126452763e"). InnerVolumeSpecName "kube-api-access-jwt5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.253425 4838 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.253461 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.253476 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwt5t\" (UniqueName: \"kubernetes.io/projected/db988bef-e07b-448a-92c3-1c126452763e-kube-api-access-jwt5t\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.253490 4838 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.253501 4838 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/db988bef-e07b-448a-92c3-1c126452763e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.253512 4838 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/db988bef-e07b-448a-92c3-1c126452763e-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.300510 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-800d-account-create-update-5qbg4" event={"ID":"344ac587-0d87-4812-9d65-18fc79eec28b","Type":"ContainerDied","Data":"43f8f10b48da146403ecf8c711a3ae1eff7b624c6ff059a376919c43883899d3"} Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.300549 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43f8f10b48da146403ecf8c711a3ae1eff7b624c6ff059a376919c43883899d3" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.300566 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-800d-account-create-update-5qbg4" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.302097 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cxm29" event={"ID":"dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c","Type":"ContainerDied","Data":"ac1b6e86fff3aa0d07d20ecf44305e56a11ecc752530a133564d2227228a60e1"} Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.302138 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac1b6e86fff3aa0d07d20ecf44305e56a11ecc752530a133564d2227228a60e1" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.302186 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cxm29" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.305935 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xql5b" event={"ID":"b524180a-e6ae-4cb3-9717-99ad184944b8","Type":"ContainerDied","Data":"f3dd70b072c20173851308a30d9f50c45502c3b8e51c26815cb8cdee353bafb9"} Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.305981 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3dd70b072c20173851308a30d9f50c45502c3b8e51c26815cb8cdee353bafb9" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.306082 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xql5b" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.312881 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kq6lf" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.312916 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kq6lf" event={"ID":"c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f","Type":"ContainerDied","Data":"1aa2760a8513ec844ebd60e8a8f77df8ff235806016e44b8143c10413aec00b2"} Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.312944 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aa2760a8513ec844ebd60e8a8f77df8ff235806016e44b8143c10413aec00b2" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.314753 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72ed-account-create-update-tx6p4" event={"ID":"a38f9117-4655-46af-8b54-01ca6b4e6553","Type":"ContainerDied","Data":"513332bd125ae9deae567dfa1b7c1a76a103366300de4751eff2c705f1b31635"} Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.314799 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="513332bd125ae9deae567dfa1b7c1a76a103366300de4751eff2c705f1b31635" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.314769 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72ed-account-create-update-tx6p4" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.316381 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f574-account-create-update-8q9qx" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.316637 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f574-account-create-update-8q9qx" event={"ID":"c2a91f97-d03b-4da9-b70c-e858e878470f","Type":"ContainerDied","Data":"2bd10724c8dd38e5549d8e57576159794f5f1b9e1d0c6988ce612015bced9ddf"} Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.316685 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bd10724c8dd38e5549d8e57576159794f5f1b9e1d0c6988ce612015bced9ddf" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.318073 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t-config-n4sbw" event={"ID":"db988bef-e07b-448a-92c3-1c126452763e","Type":"ContainerDied","Data":"7781122133ce64b614272d3d0ac1276d39f2256aadb8ce0d4624548359adfc61"} Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.318096 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7781122133ce64b614272d3d0ac1276d39f2256aadb8ce0d4624548359adfc61" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.318108 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-n4sbw" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.334054 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-sbl7t-config-n4sbw"] Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.341671 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-sbl7t-config-n4sbw"] Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.447864 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sbl7t-config-ks4q8"] Dec 07 09:25:11 crc kubenswrapper[4838]: E1207 09:25:11.448860 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2a91f97-d03b-4da9-b70c-e858e878470f" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.448993 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2a91f97-d03b-4da9-b70c-e858e878470f" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: E1207 09:25:11.449092 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38f9117-4655-46af-8b54-01ca6b4e6553" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.449178 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38f9117-4655-46af-8b54-01ca6b4e6553" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: E1207 09:25:11.449276 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.449359 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: E1207 09:25:11.449445 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db988bef-e07b-448a-92c3-1c126452763e" containerName="ovn-config" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.449529 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="db988bef-e07b-448a-92c3-1c126452763e" containerName="ovn-config" Dec 07 09:25:11 crc kubenswrapper[4838]: E1207 09:25:11.449608 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.449678 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: E1207 09:25:11.449762 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b524180a-e6ae-4cb3-9717-99ad184944b8" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.449852 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b524180a-e6ae-4cb3-9717-99ad184944b8" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: E1207 09:25:11.449939 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="344ac587-0d87-4812-9d65-18fc79eec28b" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450084 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="344ac587-0d87-4812-9d65-18fc79eec28b" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450423 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="b524180a-e6ae-4cb3-9717-99ad184944b8" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450524 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450624 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="344ac587-0d87-4812-9d65-18fc79eec28b" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450709 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2a91f97-d03b-4da9-b70c-e858e878470f" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450787 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="db988bef-e07b-448a-92c3-1c126452763e" containerName="ovn-config" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450888 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c" containerName="mariadb-database-create" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.450981 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38f9117-4655-46af-8b54-01ca6b4e6553" containerName="mariadb-account-create-update" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.451789 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.456608 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.462372 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbl7t-config-ks4q8"] Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.557251 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-additional-scripts\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.557300 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-scripts\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.557330 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.557416 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run-ovn\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.557460 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-log-ovn\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.557478 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96k9x\" (UniqueName: \"kubernetes.io/projected/b945f1fa-e964-4541-8871-18b2e4c18c4a-kube-api-access-96k9x\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.624267 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db988bef-e07b-448a-92c3-1c126452763e" path="/var/lib/kubelet/pods/db988bef-e07b-448a-92c3-1c126452763e/volumes" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.658856 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.658938 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run-ovn\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.658989 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-log-ovn\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.659005 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96k9x\" (UniqueName: \"kubernetes.io/projected/b945f1fa-e964-4541-8871-18b2e4c18c4a-kube-api-access-96k9x\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.659201 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-log-ovn\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.659210 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run-ovn\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.659217 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.659292 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-additional-scripts\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.659318 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-scripts\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.660370 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-additional-scripts\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.661745 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-scripts\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.675787 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96k9x\" (UniqueName: \"kubernetes.io/projected/b945f1fa-e964-4541-8871-18b2e4c18c4a-kube-api-access-96k9x\") pod \"ovn-controller-sbl7t-config-ks4q8\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:11 crc kubenswrapper[4838]: I1207 09:25:11.777842 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:14 crc kubenswrapper[4838]: W1207 09:25:14.892607 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb945f1fa_e964_4541_8871_18b2e4c18c4a.slice/crio-003aac5e28122099cf42426e0ee247ca405c29cee9549f403ef7554df7ce123c WatchSource:0}: Error finding container 003aac5e28122099cf42426e0ee247ca405c29cee9549f403ef7554df7ce123c: Status 404 returned error can't find the container with id 003aac5e28122099cf42426e0ee247ca405c29cee9549f403ef7554df7ce123c Dec 07 09:25:14 crc kubenswrapper[4838]: I1207 09:25:14.906091 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbl7t-config-ks4q8"] Dec 07 09:25:15 crc kubenswrapper[4838]: I1207 09:25:15.349652 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnxlk" event={"ID":"aa78bff1-c04d-42de-a3f1-7019500bd194","Type":"ContainerStarted","Data":"b3b189385765253cf03e5f3046a7f6e53e471a9f3ab49b53deef23b341ab2e4d"} Dec 07 09:25:15 crc kubenswrapper[4838]: I1207 09:25:15.354510 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t-config-ks4q8" event={"ID":"b945f1fa-e964-4541-8871-18b2e4c18c4a","Type":"ContainerStarted","Data":"33faa1be103a96fa5e0f624953e21f1c7d2519138bcd20ba9ad0c6610199ce06"} Dec 07 09:25:15 crc kubenswrapper[4838]: I1207 09:25:15.354558 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t-config-ks4q8" event={"ID":"b945f1fa-e964-4541-8871-18b2e4c18c4a","Type":"ContainerStarted","Data":"003aac5e28122099cf42426e0ee247ca405c29cee9549f403ef7554df7ce123c"} Dec 07 09:25:15 crc kubenswrapper[4838]: I1207 09:25:15.376576 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-rnxlk" podStartSLOduration=9.065617318 podStartE2EDuration="15.376549548s" podCreationTimestamp="2025-12-07 09:25:00 +0000 UTC" firstStartedPulling="2025-12-07 09:25:08.190460415 +0000 UTC m=+1124.897779432" lastFinishedPulling="2025-12-07 09:25:14.501392645 +0000 UTC m=+1131.208711662" observedRunningTime="2025-12-07 09:25:15.364419155 +0000 UTC m=+1132.071738182" watchObservedRunningTime="2025-12-07 09:25:15.376549548 +0000 UTC m=+1132.083868565" Dec 07 09:25:16 crc kubenswrapper[4838]: I1207 09:25:16.362806 4838 generic.go:334] "Generic (PLEG): container finished" podID="b945f1fa-e964-4541-8871-18b2e4c18c4a" containerID="33faa1be103a96fa5e0f624953e21f1c7d2519138bcd20ba9ad0c6610199ce06" exitCode=0 Dec 07 09:25:16 crc kubenswrapper[4838]: I1207 09:25:16.362996 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbl7t-config-ks4q8" event={"ID":"b945f1fa-e964-4541-8871-18b2e4c18c4a","Type":"ContainerDied","Data":"33faa1be103a96fa5e0f624953e21f1c7d2519138bcd20ba9ad0c6610199ce06"} Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.661110 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.758599 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run-ovn\") pod \"b945f1fa-e964-4541-8871-18b2e4c18c4a\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.758888 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96k9x\" (UniqueName: \"kubernetes.io/projected/b945f1fa-e964-4541-8871-18b2e4c18c4a-kube-api-access-96k9x\") pod \"b945f1fa-e964-4541-8871-18b2e4c18c4a\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.758723 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b945f1fa-e964-4541-8871-18b2e4c18c4a" (UID: "b945f1fa-e964-4541-8871-18b2e4c18c4a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.759711 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-additional-scripts\") pod \"b945f1fa-e964-4541-8871-18b2e4c18c4a\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.759848 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run\") pod \"b945f1fa-e964-4541-8871-18b2e4c18c4a\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.759885 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-scripts\") pod \"b945f1fa-e964-4541-8871-18b2e4c18c4a\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.759912 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-log-ovn\") pod \"b945f1fa-e964-4541-8871-18b2e4c18c4a\" (UID: \"b945f1fa-e964-4541-8871-18b2e4c18c4a\") " Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.760057 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run" (OuterVolumeSpecName: "var-run") pod "b945f1fa-e964-4541-8871-18b2e4c18c4a" (UID: "b945f1fa-e964-4541-8871-18b2e4c18c4a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.760150 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b945f1fa-e964-4541-8871-18b2e4c18c4a" (UID: "b945f1fa-e964-4541-8871-18b2e4c18c4a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.760280 4838 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.760298 4838 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-run\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.760307 4838 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b945f1fa-e964-4541-8871-18b2e4c18c4a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.760573 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b945f1fa-e964-4541-8871-18b2e4c18c4a" (UID: "b945f1fa-e964-4541-8871-18b2e4c18c4a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.760996 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-scripts" (OuterVolumeSpecName: "scripts") pod "b945f1fa-e964-4541-8871-18b2e4c18c4a" (UID: "b945f1fa-e964-4541-8871-18b2e4c18c4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.763471 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b945f1fa-e964-4541-8871-18b2e4c18c4a-kube-api-access-96k9x" (OuterVolumeSpecName: "kube-api-access-96k9x") pod "b945f1fa-e964-4541-8871-18b2e4c18c4a" (UID: "b945f1fa-e964-4541-8871-18b2e4c18c4a"). InnerVolumeSpecName "kube-api-access-96k9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.861998 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96k9x\" (UniqueName: \"kubernetes.io/projected/b945f1fa-e964-4541-8871-18b2e4c18c4a-kube-api-access-96k9x\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.862098 4838 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.862122 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b945f1fa-e964-4541-8871-18b2e4c18c4a-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.956634 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-sbl7t-config-ks4q8"] Dec 07 09:25:17 crc kubenswrapper[4838]: I1207 09:25:17.963511 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-sbl7t-config-ks4q8"] Dec 07 09:25:18 crc kubenswrapper[4838]: I1207 09:25:18.380904 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="003aac5e28122099cf42426e0ee247ca405c29cee9549f403ef7554df7ce123c" Dec 07 09:25:18 crc kubenswrapper[4838]: I1207 09:25:18.381019 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbl7t-config-ks4q8" Dec 07 09:25:18 crc kubenswrapper[4838]: I1207 09:25:18.385944 4838 generic.go:334] "Generic (PLEG): container finished" podID="eb3152f7-2ea4-4de8-ba34-038fd733682b" containerID="366f9728572c990684bf4b94aca2d5853c6b60bf195c3f16de285ed8133c3a41" exitCode=0 Dec 07 09:25:18 crc kubenswrapper[4838]: I1207 09:25:18.386004 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n2jzr" event={"ID":"eb3152f7-2ea4-4de8-ba34-038fd733682b","Type":"ContainerDied","Data":"366f9728572c990684bf4b94aca2d5853c6b60bf195c3f16de285ed8133c3a41"} Dec 07 09:25:18 crc kubenswrapper[4838]: I1207 09:25:18.388050 4838 generic.go:334] "Generic (PLEG): container finished" podID="aa78bff1-c04d-42de-a3f1-7019500bd194" containerID="b3b189385765253cf03e5f3046a7f6e53e471a9f3ab49b53deef23b341ab2e4d" exitCode=0 Dec 07 09:25:18 crc kubenswrapper[4838]: I1207 09:25:18.388106 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnxlk" event={"ID":"aa78bff1-c04d-42de-a3f1-7019500bd194","Type":"ContainerDied","Data":"b3b189385765253cf03e5f3046a7f6e53e471a9f3ab49b53deef23b341ab2e4d"} Dec 07 09:25:19 crc kubenswrapper[4838]: I1207 09:25:19.710999 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b945f1fa-e964-4541-8871-18b2e4c18c4a" path="/var/lib/kubelet/pods/b945f1fa-e964-4541-8871-18b2e4c18c4a/volumes" Dec 07 09:25:19 crc kubenswrapper[4838]: I1207 09:25:19.937631 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.014751 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n2jzr" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.101216 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-combined-ca-bundle\") pod \"aa78bff1-c04d-42de-a3f1-7019500bd194\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.101278 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnqqk\" (UniqueName: \"kubernetes.io/projected/aa78bff1-c04d-42de-a3f1-7019500bd194-kube-api-access-wnqqk\") pod \"aa78bff1-c04d-42de-a3f1-7019500bd194\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.101316 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-config-data\") pod \"aa78bff1-c04d-42de-a3f1-7019500bd194\" (UID: \"aa78bff1-c04d-42de-a3f1-7019500bd194\") " Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.108966 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa78bff1-c04d-42de-a3f1-7019500bd194-kube-api-access-wnqqk" (OuterVolumeSpecName: "kube-api-access-wnqqk") pod "aa78bff1-c04d-42de-a3f1-7019500bd194" (UID: "aa78bff1-c04d-42de-a3f1-7019500bd194"). InnerVolumeSpecName "kube-api-access-wnqqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.123662 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa78bff1-c04d-42de-a3f1-7019500bd194" (UID: "aa78bff1-c04d-42de-a3f1-7019500bd194"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.145723 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-config-data" (OuterVolumeSpecName: "config-data") pod "aa78bff1-c04d-42de-a3f1-7019500bd194" (UID: "aa78bff1-c04d-42de-a3f1-7019500bd194"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.203285 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-combined-ca-bundle\") pod \"eb3152f7-2ea4-4de8-ba34-038fd733682b\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.203382 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-config-data\") pod \"eb3152f7-2ea4-4de8-ba34-038fd733682b\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.203420 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgrgv\" (UniqueName: \"kubernetes.io/projected/eb3152f7-2ea4-4de8-ba34-038fd733682b-kube-api-access-mgrgv\") pod \"eb3152f7-2ea4-4de8-ba34-038fd733682b\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.203509 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-db-sync-config-data\") pod \"eb3152f7-2ea4-4de8-ba34-038fd733682b\" (UID: \"eb3152f7-2ea4-4de8-ba34-038fd733682b\") " Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.203859 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnqqk\" (UniqueName: \"kubernetes.io/projected/aa78bff1-c04d-42de-a3f1-7019500bd194-kube-api-access-wnqqk\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.203873 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.203882 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa78bff1-c04d-42de-a3f1-7019500bd194-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.207139 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb3152f7-2ea4-4de8-ba34-038fd733682b-kube-api-access-mgrgv" (OuterVolumeSpecName: "kube-api-access-mgrgv") pod "eb3152f7-2ea4-4de8-ba34-038fd733682b" (UID: "eb3152f7-2ea4-4de8-ba34-038fd733682b"). InnerVolumeSpecName "kube-api-access-mgrgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.207752 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eb3152f7-2ea4-4de8-ba34-038fd733682b" (UID: "eb3152f7-2ea4-4de8-ba34-038fd733682b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.223502 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb3152f7-2ea4-4de8-ba34-038fd733682b" (UID: "eb3152f7-2ea4-4de8-ba34-038fd733682b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.238988 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-config-data" (OuterVolumeSpecName: "config-data") pod "eb3152f7-2ea4-4de8-ba34-038fd733682b" (UID: "eb3152f7-2ea4-4de8-ba34-038fd733682b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.305438 4838 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.305490 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.305499 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb3152f7-2ea4-4de8-ba34-038fd733682b-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.305511 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgrgv\" (UniqueName: \"kubernetes.io/projected/eb3152f7-2ea4-4de8-ba34-038fd733682b-kube-api-access-mgrgv\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.408023 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnxlk" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.408020 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnxlk" event={"ID":"aa78bff1-c04d-42de-a3f1-7019500bd194","Type":"ContainerDied","Data":"09d2c5860ef52a06064468a846c0dab7ebbfa71faa826f5204a9003c040b971b"} Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.408468 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09d2c5860ef52a06064468a846c0dab7ebbfa71faa826f5204a9003c040b971b" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.409701 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n2jzr" event={"ID":"eb3152f7-2ea4-4de8-ba34-038fd733682b","Type":"ContainerDied","Data":"5d3dc0426a95197590360bff6fe52fe2f719b28428514bf736be03ba508fa206"} Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.409746 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d3dc0426a95197590360bff6fe52fe2f719b28428514bf736be03ba508fa206" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.409783 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n2jzr" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.705135 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-k9qq8"] Dec 07 09:25:20 crc kubenswrapper[4838]: E1207 09:25:20.705453 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b945f1fa-e964-4541-8871-18b2e4c18c4a" containerName="ovn-config" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.705466 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b945f1fa-e964-4541-8871-18b2e4c18c4a" containerName="ovn-config" Dec 07 09:25:20 crc kubenswrapper[4838]: E1207 09:25:20.705484 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb3152f7-2ea4-4de8-ba34-038fd733682b" containerName="glance-db-sync" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.705490 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb3152f7-2ea4-4de8-ba34-038fd733682b" containerName="glance-db-sync" Dec 07 09:25:20 crc kubenswrapper[4838]: E1207 09:25:20.705502 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa78bff1-c04d-42de-a3f1-7019500bd194" containerName="keystone-db-sync" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.705507 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa78bff1-c04d-42de-a3f1-7019500bd194" containerName="keystone-db-sync" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.705638 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa78bff1-c04d-42de-a3f1-7019500bd194" containerName="keystone-db-sync" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.705652 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="b945f1fa-e964-4541-8871-18b2e4c18c4a" containerName="ovn-config" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.705663 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb3152f7-2ea4-4de8-ba34-038fd733682b" containerName="glance-db-sync" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.706459 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.724687 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-k9qq8"] Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.761390 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nxmpf"] Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.762320 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.770654 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v7gbr" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.770949 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.771149 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.771312 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.772398 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.814759 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-sb\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.814853 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t87zj\" (UniqueName: \"kubernetes.io/projected/e236f70a-0f04-46f7-b596-4f01f6069769-kube-api-access-t87zj\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.814942 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-config\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.815014 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-dns-svc\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.815086 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-nb\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.910268 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nxmpf"] Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.916965 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-sb\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917041 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t87zj\" (UniqueName: \"kubernetes.io/projected/e236f70a-0f04-46f7-b596-4f01f6069769-kube-api-access-t87zj\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917081 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-config\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917127 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-scripts\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917161 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-config-data\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917186 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-dns-svc\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917216 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-credential-keys\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917237 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzqcw\" (UniqueName: \"kubernetes.io/projected/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-kube-api-access-dzqcw\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917259 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-fernet-keys\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917285 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-combined-ca-bundle\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.917313 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-nb\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.918302 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-nb\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.919008 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-sb\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.920542 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-dns-svc\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.920690 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-config\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.953287 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-p9tch"] Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.954385 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.957753 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t87zj\" (UniqueName: \"kubernetes.io/projected/e236f70a-0f04-46f7-b596-4f01f6069769-kube-api-access-t87zj\") pod \"dnsmasq-dns-8fb4c6755-k9qq8\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.961416 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.961642 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 07 09:25:20 crc kubenswrapper[4838]: I1207 09:25:20.961787 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-h7c2r" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.013437 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-p9tch"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028427 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-credential-keys\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028476 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzqcw\" (UniqueName: \"kubernetes.io/projected/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-kube-api-access-dzqcw\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028502 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-fernet-keys\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028527 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-combined-ca-bundle\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028600 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bee89ed4-9590-461e-9416-a4f0e9cad8b6-etc-machine-id\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028639 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k97k9\" (UniqueName: \"kubernetes.io/projected/bee89ed4-9590-461e-9416-a4f0e9cad8b6-kube-api-access-k97k9\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028661 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-scripts\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028714 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-config-data\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028744 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-scripts\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028766 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-db-sync-config-data\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028798 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-config-data\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.028850 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-combined-ca-bundle\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.032764 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-combined-ca-bundle\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.053529 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-credential-keys\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.055343 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-config-data\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.062873 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-fernet-keys\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.072423 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.079578 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-scripts\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.096460 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzqcw\" (UniqueName: \"kubernetes.io/projected/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-kube-api-access-dzqcw\") pod \"keystone-bootstrap-nxmpf\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.130248 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-db-sync-config-data\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.130295 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-combined-ca-bundle\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.130362 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bee89ed4-9590-461e-9416-a4f0e9cad8b6-etc-machine-id\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.130391 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k97k9\" (UniqueName: \"kubernetes.io/projected/bee89ed4-9590-461e-9416-a4f0e9cad8b6-kube-api-access-k97k9\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.130406 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-scripts\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.130441 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-config-data\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.139731 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bee89ed4-9590-461e-9416-a4f0e9cad8b6-etc-machine-id\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.152207 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-db-sync-config-data\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.152321 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-config-data\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.167300 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-scripts\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.167367 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-combined-ca-bundle\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.212302 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k97k9\" (UniqueName: \"kubernetes.io/projected/bee89ed4-9590-461e-9416-a4f0e9cad8b6-kube-api-access-k97k9\") pod \"cinder-db-sync-p9tch\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.212377 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-nw8pl"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.213544 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.217356 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.217524 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kqzfj" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.217742 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.227475 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nw8pl"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.275899 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-k9qq8"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.303115 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.305690 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.320341 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-979hq"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.320602 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.320909 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.321312 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.324690 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.325237 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vgrg9" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.325735 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.343575 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.347978 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-config-data\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348028 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8026163f-b7f8-4033-9742-79da3dde65a7-logs\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348076 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-config\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348114 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-combined-ca-bundle\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348171 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-log-httpd\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348227 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348278 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-767nx\" (UniqueName: \"kubernetes.io/projected/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-kube-api-access-767nx\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348357 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6bqh\" (UniqueName: \"kubernetes.io/projected/8026163f-b7f8-4033-9742-79da3dde65a7-kube-api-access-w6bqh\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348408 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-config-data\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348442 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-run-httpd\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348522 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348553 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-scripts\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348585 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-combined-ca-bundle\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348630 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-scripts\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.348700 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qdm5\" (UniqueName: \"kubernetes.io/projected/35d6e325-5e49-4b0c-aa41-451eb567132f-kube-api-access-8qdm5\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.354860 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-dc7m7"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.366760 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-979hq"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.366878 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.374300 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-dc7m7"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.384082 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.452666 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6bqh\" (UniqueName: \"kubernetes.io/projected/8026163f-b7f8-4033-9742-79da3dde65a7-kube-api-access-w6bqh\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.452919 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-config-data\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.453073 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-run-httpd\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.453185 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.453276 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-scripts\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.453363 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-combined-ca-bundle\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.453458 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-scripts\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.453573 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qdm5\" (UniqueName: \"kubernetes.io/projected/35d6e325-5e49-4b0c-aa41-451eb567132f-kube-api-access-8qdm5\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.457442 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-run-httpd\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.462050 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-config-data\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.462094 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8026163f-b7f8-4033-9742-79da3dde65a7-logs\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.462128 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-config\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.462162 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-combined-ca-bundle\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.462182 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-log-httpd\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.462232 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.462280 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-767nx\" (UniqueName: \"kubernetes.io/projected/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-kube-api-access-767nx\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.468317 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-log-httpd\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.470635 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.472040 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8026163f-b7f8-4033-9742-79da3dde65a7-logs\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.477209 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-scripts\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.484548 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-config\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.485583 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-combined-ca-bundle\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.486661 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p9tch" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.487689 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-combined-ca-bundle\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.491678 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-scripts\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.493236 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.503081 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6bqh\" (UniqueName: \"kubernetes.io/projected/8026163f-b7f8-4033-9742-79da3dde65a7-kube-api-access-w6bqh\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.513551 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qdm5\" (UniqueName: \"kubernetes.io/projected/35d6e325-5e49-4b0c-aa41-451eb567132f-kube-api-access-8qdm5\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.514297 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-config-data\") pod \"ceilometer-0\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.526381 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-config-data\") pod \"placement-db-sync-979hq\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.526491 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-767nx\" (UniqueName: \"kubernetes.io/projected/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-kube-api-access-767nx\") pod \"neutron-db-sync-nw8pl\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.537885 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-zhq8j"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.539106 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.544827 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kwdwg" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.545184 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.570262 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zhq8j"] Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.585331 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.597704 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt7c4\" (UniqueName: \"kubernetes.io/projected/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-kube-api-access-lt7c4\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.597786 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-db-sync-config-data\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.612704 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-dns-svc\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.612876 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-combined-ca-bundle\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.612899 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-nb\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.612928 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jrz8\" (UniqueName: \"kubernetes.io/projected/081018e6-b090-43c9-a968-bfbf5edd56ff-kube-api-access-7jrz8\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.612987 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-config\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.613077 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-sb\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.668672 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.693544 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-979hq" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.727968 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-db-sync-config-data\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.728069 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-dns-svc\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.728219 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-combined-ca-bundle\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.728251 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-nb\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.728281 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jrz8\" (UniqueName: \"kubernetes.io/projected/081018e6-b090-43c9-a968-bfbf5edd56ff-kube-api-access-7jrz8\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.728394 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-config\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.728524 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-sb\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.728650 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt7c4\" (UniqueName: \"kubernetes.io/projected/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-kube-api-access-lt7c4\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.729278 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-dns-svc\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.731484 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-nb\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.732276 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-sb\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.732307 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-config\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.737602 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-db-sync-config-data\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.742173 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-combined-ca-bundle\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.749670 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt7c4\" (UniqueName: \"kubernetes.io/projected/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-kube-api-access-lt7c4\") pod \"dnsmasq-dns-66cd6b8b67-dc7m7\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.753755 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jrz8\" (UniqueName: \"kubernetes.io/projected/081018e6-b090-43c9-a968-bfbf5edd56ff-kube-api-access-7jrz8\") pod \"barbican-db-sync-zhq8j\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:21 crc kubenswrapper[4838]: I1207 09:25:21.888457 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.009298 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.029142 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-k9qq8"] Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.113324 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nxmpf"] Dec 07 09:25:22 crc kubenswrapper[4838]: W1207 09:25:22.140449 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43e0e030_e653_4ca3_bd78_e0cd2fb3ac22.slice/crio-79316bb99625e70aa1465e667340a821338ecb7ff80c440ff2181a408e9b874a WatchSource:0}: Error finding container 79316bb99625e70aa1465e667340a821338ecb7ff80c440ff2181a408e9b874a: Status 404 returned error can't find the container with id 79316bb99625e70aa1465e667340a821338ecb7ff80c440ff2181a408e9b874a Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.274378 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-p9tch"] Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.388311 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nw8pl"] Dec 07 09:25:22 crc kubenswrapper[4838]: W1207 09:25:22.441527 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35d6e325_5e49_4b0c_aa41_451eb567132f.slice/crio-0ade4b147c0a3a4b91e2cbc4ae977411df7bf9ce5199a0b4b3b8764bce6b7f44 WatchSource:0}: Error finding container 0ade4b147c0a3a4b91e2cbc4ae977411df7bf9ce5199a0b4b3b8764bce6b7f44: Status 404 returned error can't find the container with id 0ade4b147c0a3a4b91e2cbc4ae977411df7bf9ce5199a0b4b3b8764bce6b7f44 Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.444855 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.486411 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nxmpf" event={"ID":"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22","Type":"ContainerStarted","Data":"79316bb99625e70aa1465e667340a821338ecb7ff80c440ff2181a408e9b874a"} Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.504759 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-979hq"] Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.516149 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p9tch" event={"ID":"bee89ed4-9590-461e-9416-a4f0e9cad8b6","Type":"ContainerStarted","Data":"1d333ed6eee8c552cabd98779b5996a5c669c63d74581086fcd49b4c073224ca"} Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.547063 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" event={"ID":"e236f70a-0f04-46f7-b596-4f01f6069769","Type":"ContainerStarted","Data":"0d2f4a177d2a6242e44bb752e0e448b67c7f46c97c6acb14309bf758756bcab0"} Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.564034 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nw8pl" event={"ID":"6bbd28af-439a-4841-b3b8-7c984bbd9ee7","Type":"ContainerStarted","Data":"21b40479aa8664994f29ee3a1f6b32595ba6c79f556aacb2113bf40ad6b2982f"} Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.598554 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-dc7m7"] Dec 07 09:25:22 crc kubenswrapper[4838]: I1207 09:25:22.679990 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zhq8j"] Dec 07 09:25:22 crc kubenswrapper[4838]: W1207 09:25:22.715491 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081018e6_b090_43c9_a968_bfbf5edd56ff.slice/crio-8026041a8c9d02ec1ff732ed06b7ea1b741f40d84548b7ea2884266d44b2dcef WatchSource:0}: Error finding container 8026041a8c9d02ec1ff732ed06b7ea1b741f40d84548b7ea2884266d44b2dcef: Status 404 returned error can't find the container with id 8026041a8c9d02ec1ff732ed06b7ea1b741f40d84548b7ea2884266d44b2dcef Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.584398 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerStarted","Data":"0ade4b147c0a3a4b91e2cbc4ae977411df7bf9ce5199a0b4b3b8764bce6b7f44"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.588111 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nxmpf" event={"ID":"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22","Type":"ContainerStarted","Data":"116c573ecd2d9c117ba663a2ece45430c3599ffa21d65b00d3de6339f5f00f7c"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.592088 4838 generic.go:334] "Generic (PLEG): container finished" podID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerID="1858d4dfeeb00c1f169706fd75682a2faa3a54265133f099be81dd5fc92a2d57" exitCode=0 Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.592320 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" event={"ID":"6ba38e7c-b48b-455f-bf57-6ecafc984ee2","Type":"ContainerDied","Data":"1858d4dfeeb00c1f169706fd75682a2faa3a54265133f099be81dd5fc92a2d57"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.592372 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" event={"ID":"6ba38e7c-b48b-455f-bf57-6ecafc984ee2","Type":"ContainerStarted","Data":"8a0dc7e460245fd2258f46a98dc54bfa4e2b5e86d2433255e660d5ba89a5206a"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.603647 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zhq8j" event={"ID":"081018e6-b090-43c9-a968-bfbf5edd56ff","Type":"ContainerStarted","Data":"8026041a8c9d02ec1ff732ed06b7ea1b741f40d84548b7ea2884266d44b2dcef"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.610048 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-979hq" event={"ID":"8026163f-b7f8-4033-9742-79da3dde65a7","Type":"ContainerStarted","Data":"56c8fffafab3f46dc8da0bd564c2b6af26e680b8c3b48efa4b6265bc7f09e224"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.612196 4838 generic.go:334] "Generic (PLEG): container finished" podID="e236f70a-0f04-46f7-b596-4f01f6069769" containerID="bdb1c89ea7bed84b6e0652a8787dced776d4d7dc732bafff6587ed848407bb8b" exitCode=0 Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.612369 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" event={"ID":"e236f70a-0f04-46f7-b596-4f01f6069769","Type":"ContainerDied","Data":"bdb1c89ea7bed84b6e0652a8787dced776d4d7dc732bafff6587ed848407bb8b"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.629198 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nxmpf" podStartSLOduration=3.629174576 podStartE2EDuration="3.629174576s" podCreationTimestamp="2025-12-07 09:25:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:23.606971489 +0000 UTC m=+1140.314290506" watchObservedRunningTime="2025-12-07 09:25:23.629174576 +0000 UTC m=+1140.336493593" Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.652512 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nw8pl" event={"ID":"6bbd28af-439a-4841-b3b8-7c984bbd9ee7","Type":"ContainerStarted","Data":"8988d648c2a4f2568a8b55a738c09d6afe8e990870c687e7ca003c633938a21e"} Dec 07 09:25:23 crc kubenswrapper[4838]: I1207 09:25:23.872014 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-nw8pl" podStartSLOduration=2.871997672 podStartE2EDuration="2.871997672s" podCreationTimestamp="2025-12-07 09:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:23.861203437 +0000 UTC m=+1140.568522474" watchObservedRunningTime="2025-12-07 09:25:23.871997672 +0000 UTC m=+1140.579316689" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.090355 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.193368 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.313771 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-dns-svc\") pod \"e236f70a-0f04-46f7-b596-4f01f6069769\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.313876 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t87zj\" (UniqueName: \"kubernetes.io/projected/e236f70a-0f04-46f7-b596-4f01f6069769-kube-api-access-t87zj\") pod \"e236f70a-0f04-46f7-b596-4f01f6069769\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.313989 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-config\") pod \"e236f70a-0f04-46f7-b596-4f01f6069769\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.314022 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-sb\") pod \"e236f70a-0f04-46f7-b596-4f01f6069769\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.314042 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-nb\") pod \"e236f70a-0f04-46f7-b596-4f01f6069769\" (UID: \"e236f70a-0f04-46f7-b596-4f01f6069769\") " Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.333033 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e236f70a-0f04-46f7-b596-4f01f6069769-kube-api-access-t87zj" (OuterVolumeSpecName: "kube-api-access-t87zj") pod "e236f70a-0f04-46f7-b596-4f01f6069769" (UID: "e236f70a-0f04-46f7-b596-4f01f6069769"). InnerVolumeSpecName "kube-api-access-t87zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.361201 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-config" (OuterVolumeSpecName: "config") pod "e236f70a-0f04-46f7-b596-4f01f6069769" (UID: "e236f70a-0f04-46f7-b596-4f01f6069769"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.364344 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e236f70a-0f04-46f7-b596-4f01f6069769" (UID: "e236f70a-0f04-46f7-b596-4f01f6069769"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.367303 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e236f70a-0f04-46f7-b596-4f01f6069769" (UID: "e236f70a-0f04-46f7-b596-4f01f6069769"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.376271 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e236f70a-0f04-46f7-b596-4f01f6069769" (UID: "e236f70a-0f04-46f7-b596-4f01f6069769"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.415332 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.415357 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t87zj\" (UniqueName: \"kubernetes.io/projected/e236f70a-0f04-46f7-b596-4f01f6069769-kube-api-access-t87zj\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.415367 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.415376 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.415385 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e236f70a-0f04-46f7-b596-4f01f6069769-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.494238 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.494373 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.648981 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" event={"ID":"6ba38e7c-b48b-455f-bf57-6ecafc984ee2","Type":"ContainerStarted","Data":"4d19b0be92d8b2a01518b3f4e55bb414a8d785cd20d26ae8bed7a5394c7b951b"} Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.650157 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.656834 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.657501 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fb4c6755-k9qq8" event={"ID":"e236f70a-0f04-46f7-b596-4f01f6069769","Type":"ContainerDied","Data":"0d2f4a177d2a6242e44bb752e0e448b67c7f46c97c6acb14309bf758756bcab0"} Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.657646 4838 scope.go:117] "RemoveContainer" containerID="bdb1c89ea7bed84b6e0652a8787dced776d4d7dc732bafff6587ed848407bb8b" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.684868 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" podStartSLOduration=3.684847844 podStartE2EDuration="3.684847844s" podCreationTimestamp="2025-12-07 09:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:24.668907743 +0000 UTC m=+1141.376226760" watchObservedRunningTime="2025-12-07 09:25:24.684847844 +0000 UTC m=+1141.392166861" Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.773105 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-k9qq8"] Dec 07 09:25:24 crc kubenswrapper[4838]: I1207 09:25:24.797788 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-k9qq8"] Dec 07 09:25:25 crc kubenswrapper[4838]: I1207 09:25:25.632265 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e236f70a-0f04-46f7-b596-4f01f6069769" path="/var/lib/kubelet/pods/e236f70a-0f04-46f7-b596-4f01f6069769/volumes" Dec 07 09:25:30 crc kubenswrapper[4838]: I1207 09:25:30.725118 4838 generic.go:334] "Generic (PLEG): container finished" podID="43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" containerID="116c573ecd2d9c117ba663a2ece45430c3599ffa21d65b00d3de6339f5f00f7c" exitCode=0 Dec 07 09:25:30 crc kubenswrapper[4838]: I1207 09:25:30.725460 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nxmpf" event={"ID":"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22","Type":"ContainerDied","Data":"116c573ecd2d9c117ba663a2ece45430c3599ffa21d65b00d3de6339f5f00f7c"} Dec 07 09:25:32 crc kubenswrapper[4838]: I1207 09:25:32.012102 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:25:32 crc kubenswrapper[4838]: I1207 09:25:32.098701 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-fm8dz"] Dec 07 09:25:32 crc kubenswrapper[4838]: I1207 09:25:32.099120 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="dnsmasq-dns" containerID="cri-o://9e50b3fe4418a1ee0920041fb831a5a2ad19591f0c794ec8b49af11dbed1720b" gracePeriod=10 Dec 07 09:25:32 crc kubenswrapper[4838]: I1207 09:25:32.742137 4838 generic.go:334] "Generic (PLEG): container finished" podID="3dcec780-852f-4a5a-b977-a92fd070416c" containerID="9e50b3fe4418a1ee0920041fb831a5a2ad19591f0c794ec8b49af11dbed1720b" exitCode=0 Dec 07 09:25:32 crc kubenswrapper[4838]: I1207 09:25:32.742464 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" event={"ID":"3dcec780-852f-4a5a-b977-a92fd070416c","Type":"ContainerDied","Data":"9e50b3fe4418a1ee0920041fb831a5a2ad19591f0c794ec8b49af11dbed1720b"} Dec 07 09:25:35 crc kubenswrapper[4838]: I1207 09:25:35.067946 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 07 09:25:40 crc kubenswrapper[4838]: I1207 09:25:40.067748 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 07 09:25:45 crc kubenswrapper[4838]: I1207 09:25:45.068194 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 07 09:25:45 crc kubenswrapper[4838]: I1207 09:25:45.069950 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:25:47 crc kubenswrapper[4838]: E1207 09:25:47.262585 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f" Dec 07 09:25:47 crc kubenswrapper[4838]: E1207 09:25:47.262979 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7jrz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-zhq8j_openstack(081018e6-b090-43c9-a968-bfbf5edd56ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:25:47 crc kubenswrapper[4838]: E1207 09:25:47.264126 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-zhq8j" podUID="081018e6-b090-43c9-a968-bfbf5edd56ff" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.340685 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.501338 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-combined-ca-bundle\") pod \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.501487 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-scripts\") pod \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.501521 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-credential-keys\") pod \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.502595 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-fernet-keys\") pod \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.502733 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzqcw\" (UniqueName: \"kubernetes.io/projected/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-kube-api-access-dzqcw\") pod \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.502788 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-config-data\") pod \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\" (UID: \"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22\") " Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.513754 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-scripts" (OuterVolumeSpecName: "scripts") pod "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" (UID: "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.514204 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" (UID: "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.517047 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" (UID: "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.531921 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" (UID: "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.535675 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-kube-api-access-dzqcw" (OuterVolumeSpecName: "kube-api-access-dzqcw") pod "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" (UID: "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22"). InnerVolumeSpecName "kube-api-access-dzqcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.541973 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-config-data" (OuterVolumeSpecName: "config-data") pod "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" (UID: "43e0e030-e653-4ca3-bd78-e0cd2fb3ac22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.609249 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.609292 4838 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.609304 4838 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.609313 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzqcw\" (UniqueName: \"kubernetes.io/projected/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-kube-api-access-dzqcw\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.609323 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.609332 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.874003 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nxmpf" event={"ID":"43e0e030-e653-4ca3-bd78-e0cd2fb3ac22","Type":"ContainerDied","Data":"79316bb99625e70aa1465e667340a821338ecb7ff80c440ff2181a408e9b874a"} Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.874054 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79316bb99625e70aa1465e667340a821338ecb7ff80c440ff2181a408e9b874a" Dec 07 09:25:47 crc kubenswrapper[4838]: I1207 09:25:47.874081 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nxmpf" Dec 07 09:25:47 crc kubenswrapper[4838]: E1207 09:25:47.875145 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f\\\"\"" pod="openstack/barbican-db-sync-zhq8j" podUID="081018e6-b090-43c9-a968-bfbf5edd56ff" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.435092 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nxmpf"] Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.442064 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nxmpf"] Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.535987 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-szwjk"] Dec 07 09:25:48 crc kubenswrapper[4838]: E1207 09:25:48.536335 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" containerName="keystone-bootstrap" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.536347 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" containerName="keystone-bootstrap" Dec 07 09:25:48 crc kubenswrapper[4838]: E1207 09:25:48.536368 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e236f70a-0f04-46f7-b596-4f01f6069769" containerName="init" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.536374 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e236f70a-0f04-46f7-b596-4f01f6069769" containerName="init" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.536530 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" containerName="keystone-bootstrap" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.536549 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e236f70a-0f04-46f7-b596-4f01f6069769" containerName="init" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.537094 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.538710 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.541151 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v7gbr" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.541313 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.541428 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.541533 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.547422 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-szwjk"] Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.623425 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-scripts\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.623747 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-fernet-keys\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.624696 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-899dw\" (UniqueName: \"kubernetes.io/projected/305f0d8c-8742-4318-b0be-9592574cee2f-kube-api-access-899dw\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.625869 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-config-data\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.626019 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-combined-ca-bundle\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.626193 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-credential-keys\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.727739 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-899dw\" (UniqueName: \"kubernetes.io/projected/305f0d8c-8742-4318-b0be-9592574cee2f-kube-api-access-899dw\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.727843 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-config-data\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.727862 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-combined-ca-bundle\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.727888 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-credential-keys\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.727912 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-scripts\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.727934 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-fernet-keys\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.734760 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-fernet-keys\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.734951 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-config-data\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.735434 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-scripts\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.750852 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-credential-keys\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.751629 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-899dw\" (UniqueName: \"kubernetes.io/projected/305f0d8c-8742-4318-b0be-9592574cee2f-kube-api-access-899dw\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.752556 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-combined-ca-bundle\") pod \"keystone-bootstrap-szwjk\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: E1207 09:25:48.803286 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2" Dec 07 09:25:48 crc kubenswrapper[4838]: E1207 09:25:48.803477 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k97k9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-p9tch_openstack(bee89ed4-9590-461e-9416-a4f0e9cad8b6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 09:25:48 crc kubenswrapper[4838]: E1207 09:25:48.804774 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-p9tch" podUID="bee89ed4-9590-461e-9416-a4f0e9cad8b6" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.868992 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.881999 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" event={"ID":"3dcec780-852f-4a5a-b977-a92fd070416c","Type":"ContainerDied","Data":"07bee7c5811e74f8d67dc5ab4f9afa50e2f69f99b92fe1fcd9e941e4fc53629b"} Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.882263 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07bee7c5811e74f8d67dc5ab4f9afa50e2f69f99b92fe1fcd9e941e4fc53629b" Dec 07 09:25:48 crc kubenswrapper[4838]: E1207 09:25:48.882928 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2\\\"\"" pod="openstack/cinder-db-sync-p9tch" podUID="bee89ed4-9590-461e-9416-a4f0e9cad8b6" Dec 07 09:25:48 crc kubenswrapper[4838]: I1207 09:25:48.980205 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.036685 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-dns-svc\") pod \"3dcec780-852f-4a5a-b977-a92fd070416c\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.036772 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-config\") pod \"3dcec780-852f-4a5a-b977-a92fd070416c\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.036807 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-nb\") pod \"3dcec780-852f-4a5a-b977-a92fd070416c\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.039156 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d85gt\" (UniqueName: \"kubernetes.io/projected/3dcec780-852f-4a5a-b977-a92fd070416c-kube-api-access-d85gt\") pod \"3dcec780-852f-4a5a-b977-a92fd070416c\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.039263 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-sb\") pod \"3dcec780-852f-4a5a-b977-a92fd070416c\" (UID: \"3dcec780-852f-4a5a-b977-a92fd070416c\") " Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.061164 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dcec780-852f-4a5a-b977-a92fd070416c-kube-api-access-d85gt" (OuterVolumeSpecName: "kube-api-access-d85gt") pod "3dcec780-852f-4a5a-b977-a92fd070416c" (UID: "3dcec780-852f-4a5a-b977-a92fd070416c"). InnerVolumeSpecName "kube-api-access-d85gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.132139 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3dcec780-852f-4a5a-b977-a92fd070416c" (UID: "3dcec780-852f-4a5a-b977-a92fd070416c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.139468 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-config" (OuterVolumeSpecName: "config") pod "3dcec780-852f-4a5a-b977-a92fd070416c" (UID: "3dcec780-852f-4a5a-b977-a92fd070416c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.140285 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3dcec780-852f-4a5a-b977-a92fd070416c" (UID: "3dcec780-852f-4a5a-b977-a92fd070416c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.141183 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d85gt\" (UniqueName: \"kubernetes.io/projected/3dcec780-852f-4a5a-b977-a92fd070416c-kube-api-access-d85gt\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.141198 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.141207 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.141216 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.168138 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3dcec780-852f-4a5a-b977-a92fd070416c" (UID: "3dcec780-852f-4a5a-b977-a92fd070416c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.243002 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcec780-852f-4a5a-b977-a92fd070416c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.419523 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-szwjk"] Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.632486 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43e0e030-e653-4ca3-bd78-e0cd2fb3ac22" path="/var/lib/kubelet/pods/43e0e030-e653-4ca3-bd78-e0cd2fb3ac22/volumes" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.934348 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-979hq" event={"ID":"8026163f-b7f8-4033-9742-79da3dde65a7","Type":"ContainerStarted","Data":"41ba3ec9a51af44db44c6f8f91c9297a60544b71463e959c481ce09f78437310"} Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.937966 4838 generic.go:334] "Generic (PLEG): container finished" podID="6bbd28af-439a-4841-b3b8-7c984bbd9ee7" containerID="8988d648c2a4f2568a8b55a738c09d6afe8e990870c687e7ca003c633938a21e" exitCode=0 Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.938126 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nw8pl" event={"ID":"6bbd28af-439a-4841-b3b8-7c984bbd9ee7","Type":"ContainerDied","Data":"8988d648c2a4f2568a8b55a738c09d6afe8e990870c687e7ca003c633938a21e"} Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.961084 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerStarted","Data":"06b8626a7cee0f9222810bfbb48cd1e180cad8b62851200fd3845741f476922a"} Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.964944 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-fm8dz" Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.966419 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-szwjk" event={"ID":"305f0d8c-8742-4318-b0be-9592574cee2f","Type":"ContainerStarted","Data":"e64cad8549d04bedf8f36ef7d6db865d6c6840328c438c78ffcd458105234b9b"} Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.966458 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-szwjk" event={"ID":"305f0d8c-8742-4318-b0be-9592574cee2f","Type":"ContainerStarted","Data":"a4cfb3f6b74914474a1d571f7f70fcf4bdb25e7cdef722e13167762f096d355e"} Dec 07 09:25:49 crc kubenswrapper[4838]: I1207 09:25:49.982002 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-979hq" podStartSLOduration=2.895908528 podStartE2EDuration="28.981978306s" podCreationTimestamp="2025-12-07 09:25:21 +0000 UTC" firstStartedPulling="2025-12-07 09:25:22.657376111 +0000 UTC m=+1139.364695128" lastFinishedPulling="2025-12-07 09:25:48.743445879 +0000 UTC m=+1165.450764906" observedRunningTime="2025-12-07 09:25:49.960167259 +0000 UTC m=+1166.667486296" watchObservedRunningTime="2025-12-07 09:25:49.981978306 +0000 UTC m=+1166.689297323" Dec 07 09:25:50 crc kubenswrapper[4838]: I1207 09:25:50.022086 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-fm8dz"] Dec 07 09:25:50 crc kubenswrapper[4838]: I1207 09:25:50.029850 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-fm8dz"] Dec 07 09:25:50 crc kubenswrapper[4838]: I1207 09:25:50.033800 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-szwjk" podStartSLOduration=2.033783041 podStartE2EDuration="2.033783041s" podCreationTimestamp="2025-12-07 09:25:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:50.011496961 +0000 UTC m=+1166.718815978" watchObservedRunningTime="2025-12-07 09:25:50.033783041 +0000 UTC m=+1166.741102058" Dec 07 09:25:50 crc kubenswrapper[4838]: I1207 09:25:50.974145 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerStarted","Data":"85981eb4db2fbe58d5e8624057582052e1a06bdbd874c46944a90e8fb669c38a"} Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.305306 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.405285 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-config\") pod \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.405340 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-767nx\" (UniqueName: \"kubernetes.io/projected/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-kube-api-access-767nx\") pod \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.405376 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-combined-ca-bundle\") pod \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\" (UID: \"6bbd28af-439a-4841-b3b8-7c984bbd9ee7\") " Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.429016 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-kube-api-access-767nx" (OuterVolumeSpecName: "kube-api-access-767nx") pod "6bbd28af-439a-4841-b3b8-7c984bbd9ee7" (UID: "6bbd28af-439a-4841-b3b8-7c984bbd9ee7"). InnerVolumeSpecName "kube-api-access-767nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.432158 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-config" (OuterVolumeSpecName: "config") pod "6bbd28af-439a-4841-b3b8-7c984bbd9ee7" (UID: "6bbd28af-439a-4841-b3b8-7c984bbd9ee7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.433640 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bbd28af-439a-4841-b3b8-7c984bbd9ee7" (UID: "6bbd28af-439a-4841-b3b8-7c984bbd9ee7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.506983 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.507292 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-767nx\" (UniqueName: \"kubernetes.io/projected/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-kube-api-access-767nx\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.507406 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbd28af-439a-4841-b3b8-7c984bbd9ee7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.625852 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" path="/var/lib/kubelet/pods/3dcec780-852f-4a5a-b977-a92fd070416c/volumes" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.991825 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nw8pl" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.991834 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nw8pl" event={"ID":"6bbd28af-439a-4841-b3b8-7c984bbd9ee7","Type":"ContainerDied","Data":"21b40479aa8664994f29ee3a1f6b32595ba6c79f556aacb2113bf40ad6b2982f"} Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.991901 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21b40479aa8664994f29ee3a1f6b32595ba6c79f556aacb2113bf40ad6b2982f" Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.993678 4838 generic.go:334] "Generic (PLEG): container finished" podID="8026163f-b7f8-4033-9742-79da3dde65a7" containerID="41ba3ec9a51af44db44c6f8f91c9297a60544b71463e959c481ce09f78437310" exitCode=0 Dec 07 09:25:51 crc kubenswrapper[4838]: I1207 09:25:51.993719 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-979hq" event={"ID":"8026163f-b7f8-4033-9742-79da3dde65a7","Type":"ContainerDied","Data":"41ba3ec9a51af44db44c6f8f91c9297a60544b71463e959c481ce09f78437310"} Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.232208 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-578f9b9457-kfn5d"] Dec 07 09:25:52 crc kubenswrapper[4838]: E1207 09:25:52.232548 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bbd28af-439a-4841-b3b8-7c984bbd9ee7" containerName="neutron-db-sync" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.232559 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bbd28af-439a-4841-b3b8-7c984bbd9ee7" containerName="neutron-db-sync" Dec 07 09:25:52 crc kubenswrapper[4838]: E1207 09:25:52.232575 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="dnsmasq-dns" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.232581 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="dnsmasq-dns" Dec 07 09:25:52 crc kubenswrapper[4838]: E1207 09:25:52.232592 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="init" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.232599 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="init" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.232752 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bbd28af-439a-4841-b3b8-7c984bbd9ee7" containerName="neutron-db-sync" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.232771 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dcec780-852f-4a5a-b977-a92fd070416c" containerName="dnsmasq-dns" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.240201 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.265356 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578f9b9457-kfn5d"] Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.319638 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.319688 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-config\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.319748 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-nb\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.319847 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-sb\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.319934 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld2hs\" (UniqueName: \"kubernetes.io/projected/1b96570d-092b-4374-9e9c-80c459790207-kube-api-access-ld2hs\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.416037 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-78dc87b64b-wtzk9"] Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.419952 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.420809 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-sb\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.420885 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld2hs\" (UniqueName: \"kubernetes.io/projected/1b96570d-092b-4374-9e9c-80c459790207-kube-api-access-ld2hs\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.420915 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.420933 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-config\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.420971 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-nb\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.421755 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-sb\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.422764 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-nb\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.424074 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.424255 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.424470 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kqzfj" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.425273 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.428016 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-config\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.430357 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.433785 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78dc87b64b-wtzk9"] Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.466888 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld2hs\" (UniqueName: \"kubernetes.io/projected/1b96570d-092b-4374-9e9c-80c459790207-kube-api-access-ld2hs\") pod \"dnsmasq-dns-578f9b9457-kfn5d\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.527651 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-config\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.527688 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s46d\" (UniqueName: \"kubernetes.io/projected/4224e6bd-236c-43d1-a0a1-55709ff562c7-kube-api-access-4s46d\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.527716 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-httpd-config\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.527748 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-ovndb-tls-certs\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.527768 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-combined-ca-bundle\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.564364 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.629557 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-config\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.629596 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s46d\" (UniqueName: \"kubernetes.io/projected/4224e6bd-236c-43d1-a0a1-55709ff562c7-kube-api-access-4s46d\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.629625 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-httpd-config\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.629657 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-ovndb-tls-certs\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.629671 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-combined-ca-bundle\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.633768 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-ovndb-tls-certs\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.635389 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-httpd-config\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.643585 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-combined-ca-bundle\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.653629 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-config\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.656966 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s46d\" (UniqueName: \"kubernetes.io/projected/4224e6bd-236c-43d1-a0a1-55709ff562c7-kube-api-access-4s46d\") pod \"neutron-78dc87b64b-wtzk9\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:52 crc kubenswrapper[4838]: I1207 09:25:52.789757 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.093685 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578f9b9457-kfn5d"] Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.727005 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-979hq" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.804876 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78dc87b64b-wtzk9"] Dec 07 09:25:53 crc kubenswrapper[4838]: W1207 09:25:53.816447 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4224e6bd_236c_43d1_a0a1_55709ff562c7.slice/crio-cdc146184d466ebff0ebd45fab22c01539b26694b666adbb0acb76c0c3a128aa WatchSource:0}: Error finding container cdc146184d466ebff0ebd45fab22c01539b26694b666adbb0acb76c0c3a128aa: Status 404 returned error can't find the container with id cdc146184d466ebff0ebd45fab22c01539b26694b666adbb0acb76c0c3a128aa Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.872556 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-config-data\") pod \"8026163f-b7f8-4033-9742-79da3dde65a7\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.872624 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-combined-ca-bundle\") pod \"8026163f-b7f8-4033-9742-79da3dde65a7\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.872716 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-scripts\") pod \"8026163f-b7f8-4033-9742-79da3dde65a7\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.872881 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6bqh\" (UniqueName: \"kubernetes.io/projected/8026163f-b7f8-4033-9742-79da3dde65a7-kube-api-access-w6bqh\") pod \"8026163f-b7f8-4033-9742-79da3dde65a7\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.872932 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8026163f-b7f8-4033-9742-79da3dde65a7-logs\") pod \"8026163f-b7f8-4033-9742-79da3dde65a7\" (UID: \"8026163f-b7f8-4033-9742-79da3dde65a7\") " Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.873559 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8026163f-b7f8-4033-9742-79da3dde65a7-logs" (OuterVolumeSpecName: "logs") pod "8026163f-b7f8-4033-9742-79da3dde65a7" (UID: "8026163f-b7f8-4033-9742-79da3dde65a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.879202 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8026163f-b7f8-4033-9742-79da3dde65a7-kube-api-access-w6bqh" (OuterVolumeSpecName: "kube-api-access-w6bqh") pod "8026163f-b7f8-4033-9742-79da3dde65a7" (UID: "8026163f-b7f8-4033-9742-79da3dde65a7"). InnerVolumeSpecName "kube-api-access-w6bqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.881688 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-scripts" (OuterVolumeSpecName: "scripts") pod "8026163f-b7f8-4033-9742-79da3dde65a7" (UID: "8026163f-b7f8-4033-9742-79da3dde65a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.916405 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8026163f-b7f8-4033-9742-79da3dde65a7" (UID: "8026163f-b7f8-4033-9742-79da3dde65a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.922905 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-config-data" (OuterVolumeSpecName: "config-data") pod "8026163f-b7f8-4033-9742-79da3dde65a7" (UID: "8026163f-b7f8-4033-9742-79da3dde65a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.974734 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.975136 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.975150 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8026163f-b7f8-4033-9742-79da3dde65a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.975159 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6bqh\" (UniqueName: \"kubernetes.io/projected/8026163f-b7f8-4033-9742-79da3dde65a7-kube-api-access-w6bqh\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:53 crc kubenswrapper[4838]: I1207 09:25:53.975170 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8026163f-b7f8-4033-9742-79da3dde65a7-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.027186 4838 generic.go:334] "Generic (PLEG): container finished" podID="1b96570d-092b-4374-9e9c-80c459790207" containerID="ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a" exitCode=0 Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.027263 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" event={"ID":"1b96570d-092b-4374-9e9c-80c459790207","Type":"ContainerDied","Data":"ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a"} Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.027299 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" event={"ID":"1b96570d-092b-4374-9e9c-80c459790207","Type":"ContainerStarted","Data":"b80a4e50d3c3b9b7cb257c87a157f55e232ffb0fd63679c234db61fcc160dbdb"} Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.030099 4838 generic.go:334] "Generic (PLEG): container finished" podID="305f0d8c-8742-4318-b0be-9592574cee2f" containerID="e64cad8549d04bedf8f36ef7d6db865d6c6840328c438c78ffcd458105234b9b" exitCode=0 Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.030204 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-szwjk" event={"ID":"305f0d8c-8742-4318-b0be-9592574cee2f","Type":"ContainerDied","Data":"e64cad8549d04bedf8f36ef7d6db865d6c6840328c438c78ffcd458105234b9b"} Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.034902 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc87b64b-wtzk9" event={"ID":"4224e6bd-236c-43d1-a0a1-55709ff562c7","Type":"ContainerStarted","Data":"cdc146184d466ebff0ebd45fab22c01539b26694b666adbb0acb76c0c3a128aa"} Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.039127 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-979hq" event={"ID":"8026163f-b7f8-4033-9742-79da3dde65a7","Type":"ContainerDied","Data":"56c8fffafab3f46dc8da0bd564c2b6af26e680b8c3b48efa4b6265bc7f09e224"} Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.039159 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56c8fffafab3f46dc8da0bd564c2b6af26e680b8c3b48efa4b6265bc7f09e224" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.039220 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-979hq" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.168848 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9b574bc66-nsc2v"] Dec 07 09:25:54 crc kubenswrapper[4838]: E1207 09:25:54.169264 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8026163f-b7f8-4033-9742-79da3dde65a7" containerName="placement-db-sync" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.169282 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8026163f-b7f8-4033-9742-79da3dde65a7" containerName="placement-db-sync" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.169488 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8026163f-b7f8-4033-9742-79da3dde65a7" containerName="placement-db-sync" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.171657 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.178246 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.178557 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vgrg9" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.178747 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.178935 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.179039 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.256453 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9b574bc66-nsc2v"] Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.281530 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkww8\" (UniqueName: \"kubernetes.io/projected/87509a0e-d4f9-4d24-b490-9d5e2675aa21-kube-api-access-wkww8\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.281588 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87509a0e-d4f9-4d24-b490-9d5e2675aa21-logs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.281614 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-scripts\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.281643 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-internal-tls-certs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.281669 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-config-data\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.281735 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-combined-ca-bundle\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.281756 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-public-tls-certs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.383942 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-internal-tls-certs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.384563 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-config-data\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.384664 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-combined-ca-bundle\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.384697 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-public-tls-certs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.384756 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkww8\" (UniqueName: \"kubernetes.io/projected/87509a0e-d4f9-4d24-b490-9d5e2675aa21-kube-api-access-wkww8\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.384797 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87509a0e-d4f9-4d24-b490-9d5e2675aa21-logs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.384845 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-scripts\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.386566 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87509a0e-d4f9-4d24-b490-9d5e2675aa21-logs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.389917 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-internal-tls-certs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.390036 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-scripts\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.392613 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-public-tls-certs\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.393325 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-combined-ca-bundle\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.400327 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87509a0e-d4f9-4d24-b490-9d5e2675aa21-config-data\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.404906 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkww8\" (UniqueName: \"kubernetes.io/projected/87509a0e-d4f9-4d24-b490-9d5e2675aa21-kube-api-access-wkww8\") pod \"placement-9b574bc66-nsc2v\" (UID: \"87509a0e-d4f9-4d24-b490-9d5e2675aa21\") " pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.496932 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.496980 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.497024 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.497533 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.497596 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed0d936715e5d0425ff8433d1f07f45e763ba568568132ddb5adcdb73cc5a936"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:25:54 crc kubenswrapper[4838]: I1207 09:25:54.497641 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://ed0d936715e5d0425ff8433d1f07f45e763ba568568132ddb5adcdb73cc5a936" gracePeriod=600 Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.062229 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" event={"ID":"1b96570d-092b-4374-9e9c-80c459790207","Type":"ContainerStarted","Data":"ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9"} Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.062875 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.071272 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="ed0d936715e5d0425ff8433d1f07f45e763ba568568132ddb5adcdb73cc5a936" exitCode=0 Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.071370 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"ed0d936715e5d0425ff8433d1f07f45e763ba568568132ddb5adcdb73cc5a936"} Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.071419 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9"} Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.071438 4838 scope.go:117] "RemoveContainer" containerID="b20dfe6b2c79e165940e9237c72516b3c8e4bccb48b171f998bf1346e1ad6571" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.077686 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc87b64b-wtzk9" event={"ID":"4224e6bd-236c-43d1-a0a1-55709ff562c7","Type":"ContainerStarted","Data":"d4cb4afa0c8e6650789c385c48bad71ba1e1992d1a95196c9812a375dccd5cee"} Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.077753 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc87b64b-wtzk9" event={"ID":"4224e6bd-236c-43d1-a0a1-55709ff562c7","Type":"ContainerStarted","Data":"9294585ebc68e5f3b84b5ab283abca2158c22fd3c3f8757861adc83256bacccf"} Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.078253 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.098954 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" podStartSLOduration=3.098930007 podStartE2EDuration="3.098930007s" podCreationTimestamp="2025-12-07 09:25:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:55.09214102 +0000 UTC m=+1171.799460037" watchObservedRunningTime="2025-12-07 09:25:55.098930007 +0000 UTC m=+1171.806249024" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.142507 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9b574bc66-nsc2v"] Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.153660 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-78dc87b64b-wtzk9" podStartSLOduration=3.153626987 podStartE2EDuration="3.153626987s" podCreationTimestamp="2025-12-07 09:25:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:55.126000325 +0000 UTC m=+1171.833319362" watchObservedRunningTime="2025-12-07 09:25:55.153626987 +0000 UTC m=+1171.860946004" Dec 07 09:25:55 crc kubenswrapper[4838]: W1207 09:25:55.169550 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87509a0e_d4f9_4d24_b490_9d5e2675aa21.slice/crio-610718d0d8d9c0dbcfd52fea328f5a891a2f4e1d87772a3cb4b6b00e7a414d42 WatchSource:0}: Error finding container 610718d0d8d9c0dbcfd52fea328f5a891a2f4e1d87772a3cb4b6b00e7a414d42: Status 404 returned error can't find the container with id 610718d0d8d9c0dbcfd52fea328f5a891a2f4e1d87772a3cb4b6b00e7a414d42 Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.402296 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7998cf6c8f-xbbtt"] Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.408351 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.412312 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.412500 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.427909 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7998cf6c8f-xbbtt"] Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.511802 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-public-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.511905 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-httpd-config\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.511955 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p4m4\" (UniqueName: \"kubernetes.io/projected/562c9fcd-8827-42a8-8753-d570d77656fc-kube-api-access-9p4m4\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.511983 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-combined-ca-bundle\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.512003 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-ovndb-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.512047 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-config\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.512138 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-internal-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.614103 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-internal-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.614151 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-public-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.614179 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-httpd-config\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.614222 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p4m4\" (UniqueName: \"kubernetes.io/projected/562c9fcd-8827-42a8-8753-d570d77656fc-kube-api-access-9p4m4\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.614251 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-combined-ca-bundle\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.614272 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-ovndb-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.614313 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-config\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.631223 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-config\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.649781 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-combined-ca-bundle\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.649973 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-public-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.650409 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-ovndb-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.653326 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-internal-tls-certs\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.656783 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/562c9fcd-8827-42a8-8753-d570d77656fc-httpd-config\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.660671 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p4m4\" (UniqueName: \"kubernetes.io/projected/562c9fcd-8827-42a8-8753-d570d77656fc-kube-api-access-9p4m4\") pod \"neutron-7998cf6c8f-xbbtt\" (UID: \"562c9fcd-8827-42a8-8753-d570d77656fc\") " pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:55 crc kubenswrapper[4838]: I1207 09:25:55.737327 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.006006 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.127590 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b574bc66-nsc2v" event={"ID":"87509a0e-d4f9-4d24-b490-9d5e2675aa21","Type":"ContainerStarted","Data":"c621903041aac370bb6b3a2b198b7ec811816e8357cca8f114fb904167d9c214"} Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.127631 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b574bc66-nsc2v" event={"ID":"87509a0e-d4f9-4d24-b490-9d5e2675aa21","Type":"ContainerStarted","Data":"1d599c7ef1302c9d8721cd30c3e5d425276a240bc45a981c21ee66ffbfd2d882"} Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.127642 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b574bc66-nsc2v" event={"ID":"87509a0e-d4f9-4d24-b490-9d5e2675aa21","Type":"ContainerStarted","Data":"610718d0d8d9c0dbcfd52fea328f5a891a2f4e1d87772a3cb4b6b00e7a414d42"} Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.128621 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.128642 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130093 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-szwjk" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130209 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-szwjk" event={"ID":"305f0d8c-8742-4318-b0be-9592574cee2f","Type":"ContainerDied","Data":"a4cfb3f6b74914474a1d571f7f70fcf4bdb25e7cdef722e13167762f096d355e"} Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130227 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4cfb3f6b74914474a1d571f7f70fcf4bdb25e7cdef722e13167762f096d355e" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130398 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-899dw\" (UniqueName: \"kubernetes.io/projected/305f0d8c-8742-4318-b0be-9592574cee2f-kube-api-access-899dw\") pod \"305f0d8c-8742-4318-b0be-9592574cee2f\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130460 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-combined-ca-bundle\") pod \"305f0d8c-8742-4318-b0be-9592574cee2f\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130577 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-config-data\") pod \"305f0d8c-8742-4318-b0be-9592574cee2f\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130670 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-scripts\") pod \"305f0d8c-8742-4318-b0be-9592574cee2f\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130692 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-credential-keys\") pod \"305f0d8c-8742-4318-b0be-9592574cee2f\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.130723 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-fernet-keys\") pod \"305f0d8c-8742-4318-b0be-9592574cee2f\" (UID: \"305f0d8c-8742-4318-b0be-9592574cee2f\") " Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.143067 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "305f0d8c-8742-4318-b0be-9592574cee2f" (UID: "305f0d8c-8742-4318-b0be-9592574cee2f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.159401 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-scripts" (OuterVolumeSpecName: "scripts") pod "305f0d8c-8742-4318-b0be-9592574cee2f" (UID: "305f0d8c-8742-4318-b0be-9592574cee2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.159682 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305f0d8c-8742-4318-b0be-9592574cee2f-kube-api-access-899dw" (OuterVolumeSpecName: "kube-api-access-899dw") pod "305f0d8c-8742-4318-b0be-9592574cee2f" (UID: "305f0d8c-8742-4318-b0be-9592574cee2f"). InnerVolumeSpecName "kube-api-access-899dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.160970 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "305f0d8c-8742-4318-b0be-9592574cee2f" (UID: "305f0d8c-8742-4318-b0be-9592574cee2f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.176330 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-9b574bc66-nsc2v" podStartSLOduration=2.176305479 podStartE2EDuration="2.176305479s" podCreationTimestamp="2025-12-07 09:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:25:56.160024949 +0000 UTC m=+1172.867344006" watchObservedRunningTime="2025-12-07 09:25:56.176305479 +0000 UTC m=+1172.883624496" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.206313 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-config-data" (OuterVolumeSpecName: "config-data") pod "305f0d8c-8742-4318-b0be-9592574cee2f" (UID: "305f0d8c-8742-4318-b0be-9592574cee2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.206637 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "305f0d8c-8742-4318-b0be-9592574cee2f" (UID: "305f0d8c-8742-4318-b0be-9592574cee2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.208750 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-654fd9b8c-czchm"] Dec 07 09:25:56 crc kubenswrapper[4838]: E1207 09:25:56.209176 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305f0d8c-8742-4318-b0be-9592574cee2f" containerName="keystone-bootstrap" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.209190 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="305f0d8c-8742-4318-b0be-9592574cee2f" containerName="keystone-bootstrap" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.209404 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="305f0d8c-8742-4318-b0be-9592574cee2f" containerName="keystone-bootstrap" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.210661 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.213390 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.217282 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.222620 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-654fd9b8c-czchm"] Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.236739 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-899dw\" (UniqueName: \"kubernetes.io/projected/305f0d8c-8742-4318-b0be-9592574cee2f-kube-api-access-899dw\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.236770 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.236781 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.236790 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.236800 4838 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.236808 4838 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/305f0d8c-8742-4318-b0be-9592574cee2f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338312 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-credential-keys\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338401 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-scripts\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338470 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-combined-ca-bundle\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338499 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-config-data\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338530 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-fernet-keys\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338551 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mglm\" (UniqueName: \"kubernetes.io/projected/dc215482-4e46-4b74-83da-82e55658e2d6-kube-api-access-6mglm\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338600 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-public-tls-certs\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.338670 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-internal-tls-certs\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.438359 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7998cf6c8f-xbbtt"] Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440106 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-fernet-keys\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440148 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mglm\" (UniqueName: \"kubernetes.io/projected/dc215482-4e46-4b74-83da-82e55658e2d6-kube-api-access-6mglm\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440207 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-public-tls-certs\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440272 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-internal-tls-certs\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440307 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-credential-keys\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440360 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-scripts\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440423 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-combined-ca-bundle\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.440450 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-config-data\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.446980 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-combined-ca-bundle\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.450923 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-scripts\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.454089 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-config-data\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.461699 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-credential-keys\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.463377 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mglm\" (UniqueName: \"kubernetes.io/projected/dc215482-4e46-4b74-83da-82e55658e2d6-kube-api-access-6mglm\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.466020 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-internal-tls-certs\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.466686 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-public-tls-certs\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.475543 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dc215482-4e46-4b74-83da-82e55658e2d6-fernet-keys\") pod \"keystone-654fd9b8c-czchm\" (UID: \"dc215482-4e46-4b74-83da-82e55658e2d6\") " pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:56 crc kubenswrapper[4838]: I1207 09:25:56.570850 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:25:57 crc kubenswrapper[4838]: I1207 09:25:57.102600 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-654fd9b8c-czchm"] Dec 07 09:25:57 crc kubenswrapper[4838]: I1207 09:25:57.148074 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7998cf6c8f-xbbtt" event={"ID":"562c9fcd-8827-42a8-8753-d570d77656fc","Type":"ContainerStarted","Data":"9e72c5afd8bafdcf00715d37ae2a29b29ee65cf774b1b18550221326547fdbb0"} Dec 07 09:25:57 crc kubenswrapper[4838]: I1207 09:25:57.148171 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7998cf6c8f-xbbtt" event={"ID":"562c9fcd-8827-42a8-8753-d570d77656fc","Type":"ContainerStarted","Data":"62e076b6fa7bdd0042d4d8b10641881e78a3c5efc6666551fcd1d22c07608dcb"} Dec 07 09:26:01 crc kubenswrapper[4838]: I1207 09:26:01.195235 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-654fd9b8c-czchm" event={"ID":"dc215482-4e46-4b74-83da-82e55658e2d6","Type":"ContainerStarted","Data":"4aaa878b6e7373cb24e345524a97a1f3d529940448628f07629055a81c38208c"} Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.206737 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-654fd9b8c-czchm" event={"ID":"dc215482-4e46-4b74-83da-82e55658e2d6","Type":"ContainerStarted","Data":"2946e309a91496f7b409d2539df44ba4d47fbac256479c6f1f8dcb9dc6cb4c80"} Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.207627 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.208599 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7998cf6c8f-xbbtt" event={"ID":"562c9fcd-8827-42a8-8753-d570d77656fc","Type":"ContainerStarted","Data":"c2299c53cdb718c7bd755dd40b80396210467c0b418941f001b05a3dd28ec068"} Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.208715 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.210929 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerStarted","Data":"749fab1e748fd280ad52925e0254a0eaaabadea0859b791b230bccb07b42a5e2"} Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.212836 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p9tch" event={"ID":"bee89ed4-9590-461e-9416-a4f0e9cad8b6","Type":"ContainerStarted","Data":"3828d64e8e66d54771c3226510dcf2721798e4507cf9459bc6924f8e491b76c2"} Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.213849 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zhq8j" event={"ID":"081018e6-b090-43c9-a968-bfbf5edd56ff","Type":"ContainerStarted","Data":"251722eb59ee3717debbded5a41882aeda6424519869258a500aab70e6b0786e"} Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.237699 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-654fd9b8c-czchm" podStartSLOduration=6.237679371 podStartE2EDuration="6.237679371s" podCreationTimestamp="2025-12-07 09:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:02.229928477 +0000 UTC m=+1178.937247494" watchObservedRunningTime="2025-12-07 09:26:02.237679371 +0000 UTC m=+1178.944998398" Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.289503 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-zhq8j" podStartSLOduration=2.9505726169999997 podStartE2EDuration="41.289486561s" podCreationTimestamp="2025-12-07 09:25:21 +0000 UTC" firstStartedPulling="2025-12-07 09:25:22.725526138 +0000 UTC m=+1139.432845145" lastFinishedPulling="2025-12-07 09:26:01.064440082 +0000 UTC m=+1177.771759089" observedRunningTime="2025-12-07 09:26:02.288149234 +0000 UTC m=+1178.995468261" watchObservedRunningTime="2025-12-07 09:26:02.289486561 +0000 UTC m=+1178.996805588" Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.290225 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-p9tch" podStartSLOduration=3.551516363 podStartE2EDuration="42.290218511s" podCreationTimestamp="2025-12-07 09:25:20 +0000 UTC" firstStartedPulling="2025-12-07 09:25:22.323481761 +0000 UTC m=+1139.030800778" lastFinishedPulling="2025-12-07 09:26:01.062183909 +0000 UTC m=+1177.769502926" observedRunningTime="2025-12-07 09:26:02.265333214 +0000 UTC m=+1178.972652231" watchObservedRunningTime="2025-12-07 09:26:02.290218511 +0000 UTC m=+1178.997537528" Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.567080 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.604674 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7998cf6c8f-xbbtt" podStartSLOduration=7.604642831 podStartE2EDuration="7.604642831s" podCreationTimestamp="2025-12-07 09:25:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:02.31553253 +0000 UTC m=+1179.022851567" watchObservedRunningTime="2025-12-07 09:26:02.604642831 +0000 UTC m=+1179.311961848" Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.639611 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-dc7m7"] Dec 07 09:26:02 crc kubenswrapper[4838]: I1207 09:26:02.639895 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" podUID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerName="dnsmasq-dns" containerID="cri-o://4d19b0be92d8b2a01518b3f4e55bb414a8d785cd20d26ae8bed7a5394c7b951b" gracePeriod=10 Dec 07 09:26:03 crc kubenswrapper[4838]: I1207 09:26:03.224379 4838 generic.go:334] "Generic (PLEG): container finished" podID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerID="4d19b0be92d8b2a01518b3f4e55bb414a8d785cd20d26ae8bed7a5394c7b951b" exitCode=0 Dec 07 09:26:03 crc kubenswrapper[4838]: I1207 09:26:03.224460 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" event={"ID":"6ba38e7c-b48b-455f-bf57-6ecafc984ee2","Type":"ContainerDied","Data":"4d19b0be92d8b2a01518b3f4e55bb414a8d785cd20d26ae8bed7a5394c7b951b"} Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.473677 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.607318 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-nb\") pod \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.607385 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-sb\") pod \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.607418 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-config\") pod \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.607494 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt7c4\" (UniqueName: \"kubernetes.io/projected/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-kube-api-access-lt7c4\") pod \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.607533 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-dns-svc\") pod \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\" (UID: \"6ba38e7c-b48b-455f-bf57-6ecafc984ee2\") " Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.632155 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-kube-api-access-lt7c4" (OuterVolumeSpecName: "kube-api-access-lt7c4") pod "6ba38e7c-b48b-455f-bf57-6ecafc984ee2" (UID: "6ba38e7c-b48b-455f-bf57-6ecafc984ee2"). InnerVolumeSpecName "kube-api-access-lt7c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.682191 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6ba38e7c-b48b-455f-bf57-6ecafc984ee2" (UID: "6ba38e7c-b48b-455f-bf57-6ecafc984ee2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.689197 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-config" (OuterVolumeSpecName: "config") pod "6ba38e7c-b48b-455f-bf57-6ecafc984ee2" (UID: "6ba38e7c-b48b-455f-bf57-6ecafc984ee2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.689856 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6ba38e7c-b48b-455f-bf57-6ecafc984ee2" (UID: "6ba38e7c-b48b-455f-bf57-6ecafc984ee2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.700266 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6ba38e7c-b48b-455f-bf57-6ecafc984ee2" (UID: "6ba38e7c-b48b-455f-bf57-6ecafc984ee2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.709588 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.709627 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.709643 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt7c4\" (UniqueName: \"kubernetes.io/projected/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-kube-api-access-lt7c4\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.709655 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:04 crc kubenswrapper[4838]: I1207 09:26:04.709665 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ba38e7c-b48b-455f-bf57-6ecafc984ee2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:05 crc kubenswrapper[4838]: I1207 09:26:05.256832 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" event={"ID":"6ba38e7c-b48b-455f-bf57-6ecafc984ee2","Type":"ContainerDied","Data":"8a0dc7e460245fd2258f46a98dc54bfa4e2b5e86d2433255e660d5ba89a5206a"} Dec 07 09:26:05 crc kubenswrapper[4838]: I1207 09:26:05.256888 4838 scope.go:117] "RemoveContainer" containerID="4d19b0be92d8b2a01518b3f4e55bb414a8d785cd20d26ae8bed7a5394c7b951b" Dec 07 09:26:05 crc kubenswrapper[4838]: I1207 09:26:05.256996 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-dc7m7" Dec 07 09:26:05 crc kubenswrapper[4838]: I1207 09:26:05.297258 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-dc7m7"] Dec 07 09:26:05 crc kubenswrapper[4838]: I1207 09:26:05.304453 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-dc7m7"] Dec 07 09:26:05 crc kubenswrapper[4838]: I1207 09:26:05.319025 4838 scope.go:117] "RemoveContainer" containerID="1858d4dfeeb00c1f169706fd75682a2faa3a54265133f099be81dd5fc92a2d57" Dec 07 09:26:05 crc kubenswrapper[4838]: I1207 09:26:05.625455 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" path="/var/lib/kubelet/pods/6ba38e7c-b48b-455f-bf57-6ecafc984ee2/volumes" Dec 07 09:26:06 crc kubenswrapper[4838]: I1207 09:26:06.264155 4838 generic.go:334] "Generic (PLEG): container finished" podID="081018e6-b090-43c9-a968-bfbf5edd56ff" containerID="251722eb59ee3717debbded5a41882aeda6424519869258a500aab70e6b0786e" exitCode=0 Dec 07 09:26:06 crc kubenswrapper[4838]: I1207 09:26:06.264210 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zhq8j" event={"ID":"081018e6-b090-43c9-a968-bfbf5edd56ff","Type":"ContainerDied","Data":"251722eb59ee3717debbded5a41882aeda6424519869258a500aab70e6b0786e"} Dec 07 09:26:08 crc kubenswrapper[4838]: I1207 09:26:08.288434 4838 generic.go:334] "Generic (PLEG): container finished" podID="bee89ed4-9590-461e-9416-a4f0e9cad8b6" containerID="3828d64e8e66d54771c3226510dcf2721798e4507cf9459bc6924f8e491b76c2" exitCode=0 Dec 07 09:26:08 crc kubenswrapper[4838]: I1207 09:26:08.288532 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p9tch" event={"ID":"bee89ed4-9590-461e-9416-a4f0e9cad8b6","Type":"ContainerDied","Data":"3828d64e8e66d54771c3226510dcf2721798e4507cf9459bc6924f8e491b76c2"} Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.642452 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p9tch" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.677506 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.710237 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-combined-ca-bundle\") pod \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.710280 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-config-data\") pod \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.710334 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bee89ed4-9590-461e-9416-a4f0e9cad8b6-etc-machine-id\") pod \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.710366 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-scripts\") pod \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.710382 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k97k9\" (UniqueName: \"kubernetes.io/projected/bee89ed4-9590-461e-9416-a4f0e9cad8b6-kube-api-access-k97k9\") pod \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.710405 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-db-sync-config-data\") pod \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\" (UID: \"bee89ed4-9590-461e-9416-a4f0e9cad8b6\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.711176 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bee89ed4-9590-461e-9416-a4f0e9cad8b6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bee89ed4-9590-461e-9416-a4f0e9cad8b6" (UID: "bee89ed4-9590-461e-9416-a4f0e9cad8b6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.711614 4838 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bee89ed4-9590-461e-9416-a4f0e9cad8b6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.714654 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-scripts" (OuterVolumeSpecName: "scripts") pod "bee89ed4-9590-461e-9416-a4f0e9cad8b6" (UID: "bee89ed4-9590-461e-9416-a4f0e9cad8b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.714721 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bee89ed4-9590-461e-9416-a4f0e9cad8b6-kube-api-access-k97k9" (OuterVolumeSpecName: "kube-api-access-k97k9") pod "bee89ed4-9590-461e-9416-a4f0e9cad8b6" (UID: "bee89ed4-9590-461e-9416-a4f0e9cad8b6"). InnerVolumeSpecName "kube-api-access-k97k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.715109 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bee89ed4-9590-461e-9416-a4f0e9cad8b6" (UID: "bee89ed4-9590-461e-9416-a4f0e9cad8b6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.738126 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bee89ed4-9590-461e-9416-a4f0e9cad8b6" (UID: "bee89ed4-9590-461e-9416-a4f0e9cad8b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.770698 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-config-data" (OuterVolumeSpecName: "config-data") pod "bee89ed4-9590-461e-9416-a4f0e9cad8b6" (UID: "bee89ed4-9590-461e-9416-a4f0e9cad8b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.812904 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-db-sync-config-data\") pod \"081018e6-b090-43c9-a968-bfbf5edd56ff\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.812955 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jrz8\" (UniqueName: \"kubernetes.io/projected/081018e6-b090-43c9-a968-bfbf5edd56ff-kube-api-access-7jrz8\") pod \"081018e6-b090-43c9-a968-bfbf5edd56ff\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.813038 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-combined-ca-bundle\") pod \"081018e6-b090-43c9-a968-bfbf5edd56ff\" (UID: \"081018e6-b090-43c9-a968-bfbf5edd56ff\") " Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.813988 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.814004 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.814012 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.814021 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k97k9\" (UniqueName: \"kubernetes.io/projected/bee89ed4-9590-461e-9416-a4f0e9cad8b6-kube-api-access-k97k9\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.814031 4838 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bee89ed4-9590-461e-9416-a4f0e9cad8b6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.815947 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081018e6-b090-43c9-a968-bfbf5edd56ff-kube-api-access-7jrz8" (OuterVolumeSpecName: "kube-api-access-7jrz8") pod "081018e6-b090-43c9-a968-bfbf5edd56ff" (UID: "081018e6-b090-43c9-a968-bfbf5edd56ff"). InnerVolumeSpecName "kube-api-access-7jrz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.816122 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "081018e6-b090-43c9-a968-bfbf5edd56ff" (UID: "081018e6-b090-43c9-a968-bfbf5edd56ff"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.835381 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "081018e6-b090-43c9-a968-bfbf5edd56ff" (UID: "081018e6-b090-43c9-a968-bfbf5edd56ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.915902 4838 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.915939 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jrz8\" (UniqueName: \"kubernetes.io/projected/081018e6-b090-43c9-a968-bfbf5edd56ff-kube-api-access-7jrz8\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:10 crc kubenswrapper[4838]: I1207 09:26:10.915953 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081018e6-b090-43c9-a968-bfbf5edd56ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.323565 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p9tch" event={"ID":"bee89ed4-9590-461e-9416-a4f0e9cad8b6","Type":"ContainerDied","Data":"1d333ed6eee8c552cabd98779b5996a5c669c63d74581086fcd49b4c073224ca"} Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.323612 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d333ed6eee8c552cabd98779b5996a5c669c63d74581086fcd49b4c073224ca" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.323671 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p9tch" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.327084 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zhq8j" event={"ID":"081018e6-b090-43c9-a968-bfbf5edd56ff","Type":"ContainerDied","Data":"8026041a8c9d02ec1ff732ed06b7ea1b741f40d84548b7ea2884266d44b2dcef"} Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.327248 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8026041a8c9d02ec1ff732ed06b7ea1b741f40d84548b7ea2884266d44b2dcef" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.327098 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zhq8j" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.330139 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerStarted","Data":"0b1ebeb324e37c5fd1d35030f1035cf8fca6cb89cbce03c5345e1650ebbeeb60"} Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.330280 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-central-agent" containerID="cri-o://06b8626a7cee0f9222810bfbb48cd1e180cad8b62851200fd3845741f476922a" gracePeriod=30 Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.330335 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="sg-core" containerID="cri-o://749fab1e748fd280ad52925e0254a0eaaabadea0859b791b230bccb07b42a5e2" gracePeriod=30 Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.330350 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-notification-agent" containerID="cri-o://85981eb4db2fbe58d5e8624057582052e1a06bdbd874c46944a90e8fb669c38a" gracePeriod=30 Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.330424 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="proxy-httpd" containerID="cri-o://0b1ebeb324e37c5fd1d35030f1035cf8fca6cb89cbce03c5345e1650ebbeeb60" gracePeriod=30 Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.330495 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.679290 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.610355532 podStartE2EDuration="50.679273806s" podCreationTimestamp="2025-12-07 09:25:21 +0000 UTC" firstStartedPulling="2025-12-07 09:25:22.451430688 +0000 UTC m=+1139.158749705" lastFinishedPulling="2025-12-07 09:26:10.520348962 +0000 UTC m=+1187.227667979" observedRunningTime="2025-12-07 09:26:11.371140539 +0000 UTC m=+1188.078459576" watchObservedRunningTime="2025-12-07 09:26:11.679273806 +0000 UTC m=+1188.386592823" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.947160 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:11 crc kubenswrapper[4838]: E1207 09:26:11.947583 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081018e6-b090-43c9-a968-bfbf5edd56ff" containerName="barbican-db-sync" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.947604 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="081018e6-b090-43c9-a968-bfbf5edd56ff" containerName="barbican-db-sync" Dec 07 09:26:11 crc kubenswrapper[4838]: E1207 09:26:11.947639 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee89ed4-9590-461e-9416-a4f0e9cad8b6" containerName="cinder-db-sync" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.947649 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee89ed4-9590-461e-9416-a4f0e9cad8b6" containerName="cinder-db-sync" Dec 07 09:26:11 crc kubenswrapper[4838]: E1207 09:26:11.947662 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerName="dnsmasq-dns" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.947670 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerName="dnsmasq-dns" Dec 07 09:26:11 crc kubenswrapper[4838]: E1207 09:26:11.947693 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerName="init" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.947701 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerName="init" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.960883 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="081018e6-b090-43c9-a968-bfbf5edd56ff" containerName="barbican-db-sync" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.960962 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba38e7c-b48b-455f-bf57-6ecafc984ee2" containerName="dnsmasq-dns" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.960991 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee89ed4-9590-461e-9416-a4f0e9cad8b6" containerName="cinder-db-sync" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.962003 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.965362 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.972964 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-h7c2r" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.973061 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.973244 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 07 09:26:11 crc kubenswrapper[4838]: I1207 09:26:11.974577 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.018981 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-698f6d859c-2w95m"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.020452 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.028186 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.028265 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kwdwg" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.028376 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.038235 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.038286 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.038312 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.038334 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.038375 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2qvm\" (UniqueName: \"kubernetes.io/projected/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-kube-api-access-n2qvm\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.038437 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.049225 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-d7849ff8d-sjw2m"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.051276 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.059280 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.067028 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d7849ff8d-sjw2m"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.102084 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-698f6d859c-2w95m"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144187 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2qvm\" (UniqueName: \"kubernetes.io/projected/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-kube-api-access-n2qvm\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144234 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-config-data-custom\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144275 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-combined-ca-bundle\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144309 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-combined-ca-bundle\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144337 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144359 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-config-data\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144385 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-config-data-custom\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144430 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86c6d276-ba63-41ba-aa25-10cb0994f54a-logs\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144461 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-config-data\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144493 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5tbh\" (UniqueName: \"kubernetes.io/projected/86c6d276-ba63-41ba-aa25-10cb0994f54a-kube-api-access-z5tbh\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144529 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7njg5\" (UniqueName: \"kubernetes.io/projected/2d766abd-30cf-45e4-9672-1f999276a187-kube-api-access-7njg5\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144551 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144573 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144601 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144615 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144636 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d766abd-30cf-45e4-9672-1f999276a187-logs\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.144941 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.153905 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.157479 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.160289 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.164306 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.169973 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cfc74787f-7b86z"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.171631 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.202956 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cfc74787f-7b86z"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.223177 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2qvm\" (UniqueName: \"kubernetes.io/projected/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-kube-api-access-n2qvm\") pod \"cinder-scheduler-0\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.258679 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-dns-svc\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.258963 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7njg5\" (UniqueName: \"kubernetes.io/projected/2d766abd-30cf-45e4-9672-1f999276a187-kube-api-access-7njg5\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.259251 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h7vv\" (UniqueName: \"kubernetes.io/projected/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-kube-api-access-2h7vv\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.259462 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d766abd-30cf-45e4-9672-1f999276a187-logs\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.260492 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-config-data-custom\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.260956 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-combined-ca-bundle\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.261112 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-combined-ca-bundle\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.261247 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-config\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.261334 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-sb\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.261427 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-config-data\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.261525 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-config-data-custom\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.261950 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86c6d276-ba63-41ba-aa25-10cb0994f54a-logs\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.262078 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-config-data\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.262203 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5tbh\" (UniqueName: \"kubernetes.io/projected/86c6d276-ba63-41ba-aa25-10cb0994f54a-kube-api-access-z5tbh\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.262301 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-nb\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.272471 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-combined-ca-bundle\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.274528 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-combined-ca-bundle\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.281947 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cfc74787f-7b86z"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.287148 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-config-data\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.287537 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86c6d276-ba63-41ba-aa25-10cb0994f54a-config-data-custom\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.287586 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-config-data\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.288987 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86c6d276-ba63-41ba-aa25-10cb0994f54a-logs\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.289251 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d766abd-30cf-45e4-9672-1f999276a187-logs\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.289600 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d766abd-30cf-45e4-9672-1f999276a187-config-data-custom\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.291651 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.302411 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7njg5\" (UniqueName: \"kubernetes.io/projected/2d766abd-30cf-45e4-9672-1f999276a187-kube-api-access-7njg5\") pod \"barbican-keystone-listener-d7849ff8d-sjw2m\" (UID: \"2d766abd-30cf-45e4-9672-1f999276a187\") " pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.343694 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5tbh\" (UniqueName: \"kubernetes.io/projected/86c6d276-ba63-41ba-aa25-10cb0994f54a-kube-api-access-z5tbh\") pod \"barbican-worker-698f6d859c-2w95m\" (UID: \"86c6d276-ba63-41ba-aa25-10cb0994f54a\") " pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.369537 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-698f6d859c-2w95m" Dec 07 09:26:12 crc kubenswrapper[4838]: E1207 09:26:12.371987 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-2h7vv ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" podUID="3e60bb9c-0088-4dbc-8389-ffb87d70f0f1" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.372528 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-config\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.383865 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-sb\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.385995 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-nb\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.386063 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-dns-svc\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.386155 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h7vv\" (UniqueName: \"kubernetes.io/projected/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-kube-api-access-2h7vv\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.388034 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-sb\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.388463 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-nb\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.419055 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-dns-svc\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.420274 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.426465 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h7vv\" (UniqueName: \"kubernetes.io/projected/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-kube-api-access-2h7vv\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.428166 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-config\") pod \"dnsmasq-dns-5cfc74787f-7b86z\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.437868 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-799db77f-dm9m6"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.450232 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.483478 4838 generic.go:334] "Generic (PLEG): container finished" podID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerID="0b1ebeb324e37c5fd1d35030f1035cf8fca6cb89cbce03c5345e1650ebbeeb60" exitCode=0 Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.483512 4838 generic.go:334] "Generic (PLEG): container finished" podID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerID="749fab1e748fd280ad52925e0254a0eaaabadea0859b791b230bccb07b42a5e2" exitCode=2 Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.483520 4838 generic.go:334] "Generic (PLEG): container finished" podID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerID="06b8626a7cee0f9222810bfbb48cd1e180cad8b62851200fd3845741f476922a" exitCode=0 Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.483540 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerDied","Data":"0b1ebeb324e37c5fd1d35030f1035cf8fca6cb89cbce03c5345e1650ebbeeb60"} Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.483563 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerDied","Data":"749fab1e748fd280ad52925e0254a0eaaabadea0859b791b230bccb07b42a5e2"} Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.483571 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerDied","Data":"06b8626a7cee0f9222810bfbb48cd1e180cad8b62851200fd3845741f476922a"} Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.490606 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smjj2\" (UniqueName: \"kubernetes.io/projected/f64741e5-5b47-4f9c-bbeb-f0710259803a-kube-api-access-smjj2\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.490653 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-nb\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.491503 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-sb\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.491587 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-config\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.491670 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-dns-svc\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.523123 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-799db77f-dm9m6"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.596979 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-dns-svc\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.597448 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smjj2\" (UniqueName: \"kubernetes.io/projected/f64741e5-5b47-4f9c-bbeb-f0710259803a-kube-api-access-smjj2\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.597474 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-nb\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.597549 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-sb\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.597607 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-config\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.598776 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-config\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.599635 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-nb\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.600628 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-dns-svc\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.601033 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-sb\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.602747 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-75749567f4-gxjqb"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.604378 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.610096 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.629583 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75749567f4-gxjqb"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.642177 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smjj2\" (UniqueName: \"kubernetes.io/projected/f64741e5-5b47-4f9c-bbeb-f0710259803a-kube-api-access-smjj2\") pod \"dnsmasq-dns-799db77f-dm9m6\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.657267 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.658734 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.662392 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.676828 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.690883 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699332 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a94ad2e-6c7a-4c12-843d-ca39da90f242-logs\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699365 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data-custom\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699408 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699443 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-combined-ca-bundle\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699537 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data-custom\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699568 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699583 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699604 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l98wz\" (UniqueName: \"kubernetes.io/projected/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-kube-api-access-l98wz\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699625 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt9c6\" (UniqueName: \"kubernetes.io/projected/0a94ad2e-6c7a-4c12-843d-ca39da90f242-kube-api-access-lt9c6\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699653 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a94ad2e-6c7a-4c12-843d-ca39da90f242-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699670 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-logs\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.699700 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-scripts\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802638 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data-custom\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802676 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a94ad2e-6c7a-4c12-843d-ca39da90f242-logs\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802716 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802742 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-combined-ca-bundle\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802852 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data-custom\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802883 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802898 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802919 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l98wz\" (UniqueName: \"kubernetes.io/projected/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-kube-api-access-l98wz\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802938 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt9c6\" (UniqueName: \"kubernetes.io/projected/0a94ad2e-6c7a-4c12-843d-ca39da90f242-kube-api-access-lt9c6\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802965 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a94ad2e-6c7a-4c12-843d-ca39da90f242-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.802982 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-logs\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.803004 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-scripts\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.805194 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a94ad2e-6c7a-4c12-843d-ca39da90f242-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.805617 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a94ad2e-6c7a-4c12-843d-ca39da90f242-logs\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.806096 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-logs\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.808716 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data-custom\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.809078 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-scripts\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.809602 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.809718 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data-custom\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.810227 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.810250 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-combined-ca-bundle\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.813324 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.832334 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt9c6\" (UniqueName: \"kubernetes.io/projected/0a94ad2e-6c7a-4c12-843d-ca39da90f242-kube-api-access-lt9c6\") pod \"cinder-api-0\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " pod="openstack/cinder-api-0" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.832831 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l98wz\" (UniqueName: \"kubernetes.io/projected/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-kube-api-access-l98wz\") pod \"barbican-api-75749567f4-gxjqb\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:12 crc kubenswrapper[4838]: I1207 09:26:12.994287 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.003224 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.043408 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.177756 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-698f6d859c-2w95m"] Dec 07 09:26:13 crc kubenswrapper[4838]: W1207 09:26:13.203800 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86c6d276_ba63_41ba_aa25_10cb0994f54a.slice/crio-8d907ac11e3d5a9e20100adb54dfaf0ed1423cbb19118e08372fc912ba844e5a WatchSource:0}: Error finding container 8d907ac11e3d5a9e20100adb54dfaf0ed1423cbb19118e08372fc912ba844e5a: Status 404 returned error can't find the container with id 8d907ac11e3d5a9e20100adb54dfaf0ed1423cbb19118e08372fc912ba844e5a Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.240516 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-799db77f-dm9m6"] Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.267832 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d7849ff8d-sjw2m"] Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.496131 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" event={"ID":"2d766abd-30cf-45e4-9672-1f999276a187","Type":"ContainerStarted","Data":"429b572f0d1e1d6a5f90e428a52beecf0ff658fdb2eb41b0d64761edbee62224"} Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.500281 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6242b9-8e57-4a38-b450-6ad5f9b804d6","Type":"ContainerStarted","Data":"7066c106e80818165324339f52ec22117e2e03ef3909605b48f1e1420ea65bcf"} Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.505533 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-dm9m6" event={"ID":"f64741e5-5b47-4f9c-bbeb-f0710259803a","Type":"ContainerStarted","Data":"158cd92c67c9650653995fb3ce090c8d1c7242233e5d6869d7d5de1d7c3ae1c5"} Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.507227 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.507242 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-698f6d859c-2w95m" event={"ID":"86c6d276-ba63-41ba-aa25-10cb0994f54a","Type":"ContainerStarted","Data":"8d907ac11e3d5a9e20100adb54dfaf0ed1423cbb19118e08372fc912ba844e5a"} Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.528412 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:13 crc kubenswrapper[4838]: W1207 09:26:13.579195 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fa1b1e7_72d2_4c4c_bb21_ef5c840a7c5c.slice/crio-59d17b2a422a362bc2185831e4d36402da217464d2af1b76d0d81a805b21ccf6 WatchSource:0}: Error finding container 59d17b2a422a362bc2185831e4d36402da217464d2af1b76d0d81a805b21ccf6: Status 404 returned error can't find the container with id 59d17b2a422a362bc2185831e4d36402da217464d2af1b76d0d81a805b21ccf6 Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.594297 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75749567f4-gxjqb"] Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.617998 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h7vv\" (UniqueName: \"kubernetes.io/projected/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-kube-api-access-2h7vv\") pod \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.618067 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-config\") pod \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.618107 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-dns-svc\") pod \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.618137 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-sb\") pod \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.618310 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-nb\") pod \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\" (UID: \"3e60bb9c-0088-4dbc-8389-ffb87d70f0f1\") " Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.619225 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1" (UID: "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.619235 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1" (UID: "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.619661 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1" (UID: "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.620053 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-config" (OuterVolumeSpecName: "config") pod "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1" (UID: "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.620174 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.620192 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.620205 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.620216 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.625133 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-kube-api-access-2h7vv" (OuterVolumeSpecName: "kube-api-access-2h7vv") pod "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1" (UID: "3e60bb9c-0088-4dbc-8389-ffb87d70f0f1"). InnerVolumeSpecName "kube-api-access-2h7vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.659154 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:13 crc kubenswrapper[4838]: W1207 09:26:13.664991 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a94ad2e_6c7a_4c12_843d_ca39da90f242.slice/crio-58a0685845e356e0aafba118041a0dbf2bab3b48034dba24ddeedd9c8e351dfd WatchSource:0}: Error finding container 58a0685845e356e0aafba118041a0dbf2bab3b48034dba24ddeedd9c8e351dfd: Status 404 returned error can't find the container with id 58a0685845e356e0aafba118041a0dbf2bab3b48034dba24ddeedd9c8e351dfd Dec 07 09:26:13 crc kubenswrapper[4838]: I1207 09:26:13.722121 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h7vv\" (UniqueName: \"kubernetes.io/projected/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1-kube-api-access-2h7vv\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.383517 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.520005 4838 generic.go:334] "Generic (PLEG): container finished" podID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerID="08710852290d0b1e3569aa4b8249b62b6c08951238761f5a9b47f9ab11d29223" exitCode=0 Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.520371 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-dm9m6" event={"ID":"f64741e5-5b47-4f9c-bbeb-f0710259803a","Type":"ContainerDied","Data":"08710852290d0b1e3569aa4b8249b62b6c08951238761f5a9b47f9ab11d29223"} Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.525424 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a94ad2e-6c7a-4c12-843d-ca39da90f242","Type":"ContainerStarted","Data":"58a0685845e356e0aafba118041a0dbf2bab3b48034dba24ddeedd9c8e351dfd"} Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.528767 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75749567f4-gxjqb" event={"ID":"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c","Type":"ContainerStarted","Data":"439ee46fbe938dcd548f618d4c5c8696442fd7758e234610b8685441bd07429d"} Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.528805 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75749567f4-gxjqb" event={"ID":"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c","Type":"ContainerStarted","Data":"8d1ab074842a6925aa4eb04268ce47241560a46a1d2c61e0dbb58fe003089638"} Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.528830 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75749567f4-gxjqb" event={"ID":"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c","Type":"ContainerStarted","Data":"59d17b2a422a362bc2185831e4d36402da217464d2af1b76d0d81a805b21ccf6"} Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.528869 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfc74787f-7b86z" Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.529064 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.529078 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.573733 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-75749567f4-gxjqb" podStartSLOduration=2.57371301 podStartE2EDuration="2.57371301s" podCreationTimestamp="2025-12-07 09:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:14.562609263 +0000 UTC m=+1191.269928280" watchObservedRunningTime="2025-12-07 09:26:14.57371301 +0000 UTC m=+1191.281032027" Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.708651 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cfc74787f-7b86z"] Dec 07 09:26:14 crc kubenswrapper[4838]: I1207 09:26:14.728393 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cfc74787f-7b86z"] Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.536516 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-dm9m6" event={"ID":"f64741e5-5b47-4f9c-bbeb-f0710259803a","Type":"ContainerStarted","Data":"515d7287622568f214b4608ef14abf0c9ba0c7196d8311b875c520d5baa3b029"} Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.536898 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.538584 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a94ad2e-6c7a-4c12-843d-ca39da90f242","Type":"ContainerStarted","Data":"ff998590e18368d47129a7637721425141cf3a47c53584486ce19cccdfe5147d"} Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.544780 4838 generic.go:334] "Generic (PLEG): container finished" podID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerID="85981eb4db2fbe58d5e8624057582052e1a06bdbd874c46944a90e8fb669c38a" exitCode=0 Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.544850 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerDied","Data":"85981eb4db2fbe58d5e8624057582052e1a06bdbd874c46944a90e8fb669c38a"} Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.546565 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6242b9-8e57-4a38-b450-6ad5f9b804d6","Type":"ContainerStarted","Data":"7c6a0f2185ddf0db2798a204c29a726b10a7bafe3f84d295229ae907feb14188"} Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.558122 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-799db77f-dm9m6" podStartSLOduration=3.558089095 podStartE2EDuration="3.558089095s" podCreationTimestamp="2025-12-07 09:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:15.552249433 +0000 UTC m=+1192.259568450" watchObservedRunningTime="2025-12-07 09:26:15.558089095 +0000 UTC m=+1192.265408102" Dec 07 09:26:15 crc kubenswrapper[4838]: I1207 09:26:15.668899 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e60bb9c-0088-4dbc-8389-ffb87d70f0f1" path="/var/lib/kubelet/pods/3e60bb9c-0088-4dbc-8389-ffb87d70f0f1/volumes" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.136792 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.287605 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-config-data\") pod \"35d6e325-5e49-4b0c-aa41-451eb567132f\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.287676 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-log-httpd\") pod \"35d6e325-5e49-4b0c-aa41-451eb567132f\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.287787 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-combined-ca-bundle\") pod \"35d6e325-5e49-4b0c-aa41-451eb567132f\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.287854 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qdm5\" (UniqueName: \"kubernetes.io/projected/35d6e325-5e49-4b0c-aa41-451eb567132f-kube-api-access-8qdm5\") pod \"35d6e325-5e49-4b0c-aa41-451eb567132f\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.287899 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-run-httpd\") pod \"35d6e325-5e49-4b0c-aa41-451eb567132f\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.287985 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-sg-core-conf-yaml\") pod \"35d6e325-5e49-4b0c-aa41-451eb567132f\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.288025 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-scripts\") pod \"35d6e325-5e49-4b0c-aa41-451eb567132f\" (UID: \"35d6e325-5e49-4b0c-aa41-451eb567132f\") " Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.288588 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "35d6e325-5e49-4b0c-aa41-451eb567132f" (UID: "35d6e325-5e49-4b0c-aa41-451eb567132f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.289036 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "35d6e325-5e49-4b0c-aa41-451eb567132f" (UID: "35d6e325-5e49-4b0c-aa41-451eb567132f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.294724 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d6e325-5e49-4b0c-aa41-451eb567132f-kube-api-access-8qdm5" (OuterVolumeSpecName: "kube-api-access-8qdm5") pod "35d6e325-5e49-4b0c-aa41-451eb567132f" (UID: "35d6e325-5e49-4b0c-aa41-451eb567132f"). InnerVolumeSpecName "kube-api-access-8qdm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.298560 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-scripts" (OuterVolumeSpecName: "scripts") pod "35d6e325-5e49-4b0c-aa41-451eb567132f" (UID: "35d6e325-5e49-4b0c-aa41-451eb567132f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.321304 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "35d6e325-5e49-4b0c-aa41-451eb567132f" (UID: "35d6e325-5e49-4b0c-aa41-451eb567132f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.398199 4838 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.398235 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.398247 4838 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.398256 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qdm5\" (UniqueName: \"kubernetes.io/projected/35d6e325-5e49-4b0c-aa41-451eb567132f-kube-api-access-8qdm5\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.398270 4838 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35d6e325-5e49-4b0c-aa41-451eb567132f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.405687 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35d6e325-5e49-4b0c-aa41-451eb567132f" (UID: "35d6e325-5e49-4b0c-aa41-451eb567132f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.428729 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-config-data" (OuterVolumeSpecName: "config-data") pod "35d6e325-5e49-4b0c-aa41-451eb567132f" (UID: "35d6e325-5e49-4b0c-aa41-451eb567132f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.499859 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.500069 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35d6e325-5e49-4b0c-aa41-451eb567132f-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.564052 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35d6e325-5e49-4b0c-aa41-451eb567132f","Type":"ContainerDied","Data":"0ade4b147c0a3a4b91e2cbc4ae977411df7bf9ce5199a0b4b3b8764bce6b7f44"} Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.564089 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.564135 4838 scope.go:117] "RemoveContainer" containerID="0b1ebeb324e37c5fd1d35030f1035cf8fca6cb89cbce03c5345e1650ebbeeb60" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.570852 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-698f6d859c-2w95m" event={"ID":"86c6d276-ba63-41ba-aa25-10cb0994f54a","Type":"ContainerStarted","Data":"728dd286d30777dd32065c27cd8fcd07ec933b08263ca651c592afe5c3e70ba1"} Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.570895 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-698f6d859c-2w95m" event={"ID":"86c6d276-ba63-41ba-aa25-10cb0994f54a","Type":"ContainerStarted","Data":"c0aef51a16b960fd602c9d346e9c3fb8902eb59f2631fa6449fcf9950dc4f636"} Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.575046 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" event={"ID":"2d766abd-30cf-45e4-9672-1f999276a187","Type":"ContainerStarted","Data":"708bd0753f5c406470b8bfb9414a8850fedf8ffb1a129b6fedf1d1d9cc1f12f9"} Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.575089 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" event={"ID":"2d766abd-30cf-45e4-9672-1f999276a187","Type":"ContainerStarted","Data":"ef096acc623ab66e7022cfe5e1701a0a8000578696d943bcc466e682def31764"} Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.595744 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-698f6d859c-2w95m" podStartSLOduration=3.01656693 podStartE2EDuration="5.59571853s" podCreationTimestamp="2025-12-07 09:26:11 +0000 UTC" firstStartedPulling="2025-12-07 09:26:13.207004871 +0000 UTC m=+1189.914323898" lastFinishedPulling="2025-12-07 09:26:15.786156481 +0000 UTC m=+1192.493475498" observedRunningTime="2025-12-07 09:26:16.588310135 +0000 UTC m=+1193.295629162" watchObservedRunningTime="2025-12-07 09:26:16.59571853 +0000 UTC m=+1193.303037547" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.599713 4838 scope.go:117] "RemoveContainer" containerID="749fab1e748fd280ad52925e0254a0eaaabadea0859b791b230bccb07b42a5e2" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.627029 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-d7849ff8d-sjw2m" podStartSLOduration=3.117756173 podStartE2EDuration="5.626926221s" podCreationTimestamp="2025-12-07 09:26:11 +0000 UTC" firstStartedPulling="2025-12-07 09:26:13.307747112 +0000 UTC m=+1190.015066129" lastFinishedPulling="2025-12-07 09:26:15.81691716 +0000 UTC m=+1192.524236177" observedRunningTime="2025-12-07 09:26:16.610088156 +0000 UTC m=+1193.317407173" watchObservedRunningTime="2025-12-07 09:26:16.626926221 +0000 UTC m=+1193.334245238" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.748899 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.763063 4838 scope.go:117] "RemoveContainer" containerID="85981eb4db2fbe58d5e8624057582052e1a06bdbd874c46944a90e8fb669c38a" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.781826 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.795274 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:16 crc kubenswrapper[4838]: E1207 09:26:16.795767 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-notification-agent" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.795792 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-notification-agent" Dec 07 09:26:16 crc kubenswrapper[4838]: E1207 09:26:16.795809 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-central-agent" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.795841 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-central-agent" Dec 07 09:26:16 crc kubenswrapper[4838]: E1207 09:26:16.795864 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="proxy-httpd" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.795872 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="proxy-httpd" Dec 07 09:26:16 crc kubenswrapper[4838]: E1207 09:26:16.795887 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="sg-core" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.795895 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="sg-core" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.796249 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-notification-agent" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.796272 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="proxy-httpd" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.796291 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="ceilometer-central-agent" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.796299 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" containerName="sg-core" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.798043 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.799294 4838 scope.go:117] "RemoveContainer" containerID="06b8626a7cee0f9222810bfbb48cd1e180cad8b62851200fd3845741f476922a" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.802324 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.812401 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.812651 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.925910 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-scripts\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.926549 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-log-httpd\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.926781 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gmxj\" (UniqueName: \"kubernetes.io/projected/a0382eb7-0fc9-44ff-a191-3b9b15526127-kube-api-access-2gmxj\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.926846 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-config-data\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.927064 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.927154 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-run-httpd\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:16 crc kubenswrapper[4838]: I1207 09:26:16.927181 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.028454 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-log-httpd\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.028554 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gmxj\" (UniqueName: \"kubernetes.io/projected/a0382eb7-0fc9-44ff-a191-3b9b15526127-kube-api-access-2gmxj\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.028583 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-config-data\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.028629 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.028653 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-run-httpd\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.028674 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.028744 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-scripts\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.029149 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-log-httpd\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.029447 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-run-httpd\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.032778 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-scripts\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.033010 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.035557 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-config-data\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.042369 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.049431 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gmxj\" (UniqueName: \"kubernetes.io/projected/a0382eb7-0fc9-44ff-a191-3b9b15526127-kube-api-access-2gmxj\") pod \"ceilometer-0\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.139301 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.585441 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a94ad2e-6c7a-4c12-843d-ca39da90f242","Type":"ContainerStarted","Data":"512f84a9d74359f3d555800b05dc644c65eaf717a4dddc2f3bfb5d00b4777336"} Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.585899 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api-log" containerID="cri-o://ff998590e18368d47129a7637721425141cf3a47c53584486ce19cccdfe5147d" gracePeriod=30 Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.586147 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.586394 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api" containerID="cri-o://512f84a9d74359f3d555800b05dc644c65eaf717a4dddc2f3bfb5d00b4777336" gracePeriod=30 Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.592857 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6242b9-8e57-4a38-b450-6ad5f9b804d6","Type":"ContainerStarted","Data":"f7eea783ea7f3e5918b2b537f29718db4c868b2fe5999776c60bd87363840a94"} Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.625112 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.625093786 podStartE2EDuration="5.625093786s" podCreationTimestamp="2025-12-07 09:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:17.618261127 +0000 UTC m=+1194.325580164" watchObservedRunningTime="2025-12-07 09:26:17.625093786 +0000 UTC m=+1194.332412793" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.628566 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d6e325-5e49-4b0c-aa41-451eb567132f" path="/var/lib/kubelet/pods/35d6e325-5e49-4b0c-aa41-451eb567132f/volumes" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.646622 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.651543881 podStartE2EDuration="6.64660654s" podCreationTimestamp="2025-12-07 09:26:11 +0000 UTC" firstStartedPulling="2025-12-07 09:26:13.077260599 +0000 UTC m=+1189.784579616" lastFinishedPulling="2025-12-07 09:26:14.072323258 +0000 UTC m=+1190.779642275" observedRunningTime="2025-12-07 09:26:17.638412904 +0000 UTC m=+1194.345731921" watchObservedRunningTime="2025-12-07 09:26:17.64660654 +0000 UTC m=+1194.353925557" Dec 07 09:26:17 crc kubenswrapper[4838]: I1207 09:26:17.772129 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:17 crc kubenswrapper[4838]: W1207 09:26:17.774521 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0382eb7_0fc9_44ff_a191_3b9b15526127.slice/crio-109b28ee2c45ba145bd9a71926e1e98c5143464e4594ef2a70506dce697dc427 WatchSource:0}: Error finding container 109b28ee2c45ba145bd9a71926e1e98c5143464e4594ef2a70506dce697dc427: Status 404 returned error can't find the container with id 109b28ee2c45ba145bd9a71926e1e98c5143464e4594ef2a70506dce697dc427 Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.602418 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerStarted","Data":"109b28ee2c45ba145bd9a71926e1e98c5143464e4594ef2a70506dce697dc427"} Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.605265 4838 generic.go:334] "Generic (PLEG): container finished" podID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerID="512f84a9d74359f3d555800b05dc644c65eaf717a4dddc2f3bfb5d00b4777336" exitCode=0 Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.605311 4838 generic.go:334] "Generic (PLEG): container finished" podID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerID="ff998590e18368d47129a7637721425141cf3a47c53584486ce19cccdfe5147d" exitCode=143 Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.606086 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a94ad2e-6c7a-4c12-843d-ca39da90f242","Type":"ContainerDied","Data":"512f84a9d74359f3d555800b05dc644c65eaf717a4dddc2f3bfb5d00b4777336"} Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.606120 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a94ad2e-6c7a-4c12-843d-ca39da90f242","Type":"ContainerDied","Data":"ff998590e18368d47129a7637721425141cf3a47c53584486ce19cccdfe5147d"} Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.659147 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-59c4fb88bd-m7xbq"] Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.660795 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.663549 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.663700 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.695893 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-59c4fb88bd-m7xbq"] Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.859226 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-public-tls-certs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.859309 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-internal-tls-certs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.859334 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279b7609-3731-4b2d-a74f-64b591e53925-logs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.860357 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-config-data\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.860541 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-config-data-custom\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.860647 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-combined-ca-bundle\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.860747 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pvpw\" (UniqueName: \"kubernetes.io/projected/279b7609-3731-4b2d-a74f-64b591e53925-kube-api-access-4pvpw\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.962714 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-config-data-custom\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.962786 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-combined-ca-bundle\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.962843 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pvpw\" (UniqueName: \"kubernetes.io/projected/279b7609-3731-4b2d-a74f-64b591e53925-kube-api-access-4pvpw\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.962912 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-public-tls-certs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.962955 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-internal-tls-certs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.962985 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279b7609-3731-4b2d-a74f-64b591e53925-logs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.963105 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-config-data\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.963685 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279b7609-3731-4b2d-a74f-64b591e53925-logs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.971848 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-internal-tls-certs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.972512 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-combined-ca-bundle\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.973256 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-config-data-custom\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.974487 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-public-tls-certs\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.980702 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279b7609-3731-4b2d-a74f-64b591e53925-config-data\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:18 crc kubenswrapper[4838]: I1207 09:26:18.996710 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pvpw\" (UniqueName: \"kubernetes.io/projected/279b7609-3731-4b2d-a74f-64b591e53925-kube-api-access-4pvpw\") pod \"barbican-api-59c4fb88bd-m7xbq\" (UID: \"279b7609-3731-4b2d-a74f-64b591e53925\") " pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:19 crc kubenswrapper[4838]: I1207 09:26:19.285131 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:19 crc kubenswrapper[4838]: I1207 09:26:19.904913 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-59c4fb88bd-m7xbq"] Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.135204 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240137 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-combined-ca-bundle\") pod \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240209 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a94ad2e-6c7a-4c12-843d-ca39da90f242-logs\") pod \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240232 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-scripts\") pod \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240278 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt9c6\" (UniqueName: \"kubernetes.io/projected/0a94ad2e-6c7a-4c12-843d-ca39da90f242-kube-api-access-lt9c6\") pod \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240317 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data\") pod \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240350 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data-custom\") pod \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240408 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a94ad2e-6c7a-4c12-843d-ca39da90f242-etc-machine-id\") pod \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\" (UID: \"0a94ad2e-6c7a-4c12-843d-ca39da90f242\") " Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.240793 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a94ad2e-6c7a-4c12-843d-ca39da90f242-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0a94ad2e-6c7a-4c12-843d-ca39da90f242" (UID: "0a94ad2e-6c7a-4c12-843d-ca39da90f242"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.243186 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a94ad2e-6c7a-4c12-843d-ca39da90f242-logs" (OuterVolumeSpecName: "logs") pod "0a94ad2e-6c7a-4c12-843d-ca39da90f242" (UID: "0a94ad2e-6c7a-4c12-843d-ca39da90f242"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.256733 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-scripts" (OuterVolumeSpecName: "scripts") pod "0a94ad2e-6c7a-4c12-843d-ca39da90f242" (UID: "0a94ad2e-6c7a-4c12-843d-ca39da90f242"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.257883 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0a94ad2e-6c7a-4c12-843d-ca39da90f242" (UID: "0a94ad2e-6c7a-4c12-843d-ca39da90f242"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.264078 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a94ad2e-6c7a-4c12-843d-ca39da90f242-kube-api-access-lt9c6" (OuterVolumeSpecName: "kube-api-access-lt9c6") pod "0a94ad2e-6c7a-4c12-843d-ca39da90f242" (UID: "0a94ad2e-6c7a-4c12-843d-ca39da90f242"). InnerVolumeSpecName "kube-api-access-lt9c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.288717 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a94ad2e-6c7a-4c12-843d-ca39da90f242" (UID: "0a94ad2e-6c7a-4c12-843d-ca39da90f242"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.323061 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data" (OuterVolumeSpecName: "config-data") pod "0a94ad2e-6c7a-4c12-843d-ca39da90f242" (UID: "0a94ad2e-6c7a-4c12-843d-ca39da90f242"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.342955 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.343140 4838 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.343243 4838 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a94ad2e-6c7a-4c12-843d-ca39da90f242-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.343315 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.343385 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a94ad2e-6c7a-4c12-843d-ca39da90f242-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.343453 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a94ad2e-6c7a-4c12-843d-ca39da90f242-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.343536 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt9c6\" (UniqueName: \"kubernetes.io/projected/0a94ad2e-6c7a-4c12-843d-ca39da90f242-kube-api-access-lt9c6\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.654134 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59c4fb88bd-m7xbq" event={"ID":"279b7609-3731-4b2d-a74f-64b591e53925","Type":"ContainerStarted","Data":"befbd424eb7e15c5491c055c2ff80f8d13970738c05dcaef7cc5ca292365048e"} Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.655321 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59c4fb88bd-m7xbq" event={"ID":"279b7609-3731-4b2d-a74f-64b591e53925","Type":"ContainerStarted","Data":"c69fa899cc0f327eb321a813f14d4669c4685dfe86a173031358a30eeef767a1"} Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.655392 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.655452 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.655506 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59c4fb88bd-m7xbq" event={"ID":"279b7609-3731-4b2d-a74f-64b591e53925","Type":"ContainerStarted","Data":"fce1d3aa882850378e0fb1b10afd3889b8801fb3993d86f46c759975a3376f2b"} Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.666665 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerStarted","Data":"553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f"} Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.666739 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerStarted","Data":"b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e"} Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.688030 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-59c4fb88bd-m7xbq" podStartSLOduration=2.688005821 podStartE2EDuration="2.688005821s" podCreationTimestamp="2025-12-07 09:26:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:20.687551418 +0000 UTC m=+1197.394870435" watchObservedRunningTime="2025-12-07 09:26:20.688005821 +0000 UTC m=+1197.395324828" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.699139 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a94ad2e-6c7a-4c12-843d-ca39da90f242","Type":"ContainerDied","Data":"58a0685845e356e0aafba118041a0dbf2bab3b48034dba24ddeedd9c8e351dfd"} Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.699219 4838 scope.go:117] "RemoveContainer" containerID="512f84a9d74359f3d555800b05dc644c65eaf717a4dddc2f3bfb5d00b4777336" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.699252 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.742714 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.749196 4838 scope.go:117] "RemoveContainer" containerID="ff998590e18368d47129a7637721425141cf3a47c53584486ce19cccdfe5147d" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.759608 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.793810 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:20 crc kubenswrapper[4838]: E1207 09:26:20.794319 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api-log" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.794402 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api-log" Dec 07 09:26:20 crc kubenswrapper[4838]: E1207 09:26:20.794476 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.794532 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.794736 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api-log" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.794796 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" containerName="cinder-api" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.796629 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.799196 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.799445 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.799701 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.803876 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.854233 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7ptm\" (UniqueName: \"kubernetes.io/projected/7e943cd3-c74f-4317-951f-5d1696ee7ffd-kube-api-access-l7ptm\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.854833 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-scripts\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.854928 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e943cd3-c74f-4317-951f-5d1696ee7ffd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.855000 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e943cd3-c74f-4317-951f-5d1696ee7ffd-logs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.855086 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-config-data-custom\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.855161 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-config-data\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.855260 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.855547 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.855706 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.957766 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-config-data-custom\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.959334 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-config-data\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.959785 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.960075 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.960187 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.960304 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7ptm\" (UniqueName: \"kubernetes.io/projected/7e943cd3-c74f-4317-951f-5d1696ee7ffd-kube-api-access-l7ptm\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.960384 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-scripts\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.960454 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e943cd3-c74f-4317-951f-5d1696ee7ffd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.960539 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e943cd3-c74f-4317-951f-5d1696ee7ffd-logs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.961204 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e943cd3-c74f-4317-951f-5d1696ee7ffd-logs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.961242 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e943cd3-c74f-4317-951f-5d1696ee7ffd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.963965 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-config-data-custom\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.965417 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-config-data\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.972183 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-scripts\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.979054 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:20 crc kubenswrapper[4838]: I1207 09:26:20.992113 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:21 crc kubenswrapper[4838]: I1207 09:26:20.994478 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e943cd3-c74f-4317-951f-5d1696ee7ffd-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:21 crc kubenswrapper[4838]: I1207 09:26:20.997173 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7ptm\" (UniqueName: \"kubernetes.io/projected/7e943cd3-c74f-4317-951f-5d1696ee7ffd-kube-api-access-l7ptm\") pod \"cinder-api-0\" (UID: \"7e943cd3-c74f-4317-951f-5d1696ee7ffd\") " pod="openstack/cinder-api-0" Dec 07 09:26:21 crc kubenswrapper[4838]: I1207 09:26:21.114862 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 09:26:21 crc kubenswrapper[4838]: I1207 09:26:21.608993 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 09:26:21 crc kubenswrapper[4838]: W1207 09:26:21.620416 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e943cd3_c74f_4317_951f_5d1696ee7ffd.slice/crio-dc5c8ccbab31bb7a0a9231c7ecd4fb28699e462286e4fb3ef9aea304cdf1d9ca WatchSource:0}: Error finding container dc5c8ccbab31bb7a0a9231c7ecd4fb28699e462286e4fb3ef9aea304cdf1d9ca: Status 404 returned error can't find the container with id dc5c8ccbab31bb7a0a9231c7ecd4fb28699e462286e4fb3ef9aea304cdf1d9ca Dec 07 09:26:21 crc kubenswrapper[4838]: I1207 09:26:21.624945 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a94ad2e-6c7a-4c12-843d-ca39da90f242" path="/var/lib/kubelet/pods/0a94ad2e-6c7a-4c12-843d-ca39da90f242/volumes" Dec 07 09:26:21 crc kubenswrapper[4838]: I1207 09:26:21.707608 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e943cd3-c74f-4317-951f-5d1696ee7ffd","Type":"ContainerStarted","Data":"dc5c8ccbab31bb7a0a9231c7ecd4fb28699e462286e4fb3ef9aea304cdf1d9ca"} Dec 07 09:26:21 crc kubenswrapper[4838]: I1207 09:26:21.709969 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerStarted","Data":"f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737"} Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.292403 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.512504 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.678963 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.756748 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e943cd3-c74f-4317-951f-5d1696ee7ffd","Type":"ContainerStarted","Data":"d44656df267c63c58542ac9bd285596e119edfae06c0f755104f1f9cf7d553db"} Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.758297 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578f9b9457-kfn5d"] Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.759192 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" podUID="1b96570d-092b-4374-9e9c-80c459790207" containerName="dnsmasq-dns" containerID="cri-o://ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9" gracePeriod=10 Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.830771 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:26:22 crc kubenswrapper[4838]: I1207 09:26:22.836987 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.388074 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.511129 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc\") pod \"1b96570d-092b-4374-9e9c-80c459790207\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.511535 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-sb\") pod \"1b96570d-092b-4374-9e9c-80c459790207\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.511595 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld2hs\" (UniqueName: \"kubernetes.io/projected/1b96570d-092b-4374-9e9c-80c459790207-kube-api-access-ld2hs\") pod \"1b96570d-092b-4374-9e9c-80c459790207\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.511653 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-config\") pod \"1b96570d-092b-4374-9e9c-80c459790207\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.511765 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-nb\") pod \"1b96570d-092b-4374-9e9c-80c459790207\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.529837 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b96570d-092b-4374-9e9c-80c459790207-kube-api-access-ld2hs" (OuterVolumeSpecName: "kube-api-access-ld2hs") pod "1b96570d-092b-4374-9e9c-80c459790207" (UID: "1b96570d-092b-4374-9e9c-80c459790207"). InnerVolumeSpecName "kube-api-access-ld2hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.601333 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b96570d-092b-4374-9e9c-80c459790207" (UID: "1b96570d-092b-4374-9e9c-80c459790207"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.613724 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.613757 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld2hs\" (UniqueName: \"kubernetes.io/projected/1b96570d-092b-4374-9e9c-80c459790207-kube-api-access-ld2hs\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.648256 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-config" (OuterVolumeSpecName: "config") pod "1b96570d-092b-4374-9e9c-80c459790207" (UID: "1b96570d-092b-4374-9e9c-80c459790207"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:23 crc kubenswrapper[4838]: E1207 09:26:23.670179 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc podName:1b96570d-092b-4374-9e9c-80c459790207 nodeName:}" failed. No retries permitted until 2025-12-07 09:26:24.170157057 +0000 UTC m=+1200.877476074 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc") pod "1b96570d-092b-4374-9e9c-80c459790207" (UID: "1b96570d-092b-4374-9e9c-80c459790207") : error deleting /var/lib/kubelet/pods/1b96570d-092b-4374-9e9c-80c459790207/volume-subpaths: remove /var/lib/kubelet/pods/1b96570d-092b-4374-9e9c-80c459790207/volume-subpaths: no such file or directory Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.670514 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b96570d-092b-4374-9e9c-80c459790207" (UID: "1b96570d-092b-4374-9e9c-80c459790207"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.715092 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.715114 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.765721 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerStarted","Data":"341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145"} Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.765892 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.767664 4838 generic.go:334] "Generic (PLEG): container finished" podID="1b96570d-092b-4374-9e9c-80c459790207" containerID="ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9" exitCode=0 Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.767714 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" event={"ID":"1b96570d-092b-4374-9e9c-80c459790207","Type":"ContainerDied","Data":"ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9"} Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.767745 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" event={"ID":"1b96570d-092b-4374-9e9c-80c459790207","Type":"ContainerDied","Data":"b80a4e50d3c3b9b7cb257c87a157f55e232ffb0fd63679c234db61fcc160dbdb"} Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.767769 4838 scope.go:117] "RemoveContainer" containerID="ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.767921 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578f9b9457-kfn5d" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.775227 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e943cd3-c74f-4317-951f-5d1696ee7ffd","Type":"ContainerStarted","Data":"b6c4d7ca13d9f76db3aab96aa1cc5028d6a62440d995eb9923c8031ed01e6f6b"} Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.775343 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="cinder-scheduler" containerID="cri-o://7c6a0f2185ddf0db2798a204c29a726b10a7bafe3f84d295229ae907feb14188" gracePeriod=30 Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.775707 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="probe" containerID="cri-o://f7eea783ea7f3e5918b2b537f29718db4c868b2fe5999776c60bd87363840a94" gracePeriod=30 Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.791019 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.949341063 podStartE2EDuration="7.791005043s" podCreationTimestamp="2025-12-07 09:26:16 +0000 UTC" firstStartedPulling="2025-12-07 09:26:17.777959806 +0000 UTC m=+1194.485278823" lastFinishedPulling="2025-12-07 09:26:22.619623786 +0000 UTC m=+1199.326942803" observedRunningTime="2025-12-07 09:26:23.786624782 +0000 UTC m=+1200.493943799" watchObservedRunningTime="2025-12-07 09:26:23.791005043 +0000 UTC m=+1200.498324060" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.818341 4838 scope.go:117] "RemoveContainer" containerID="ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.866194 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.8661723180000003 podStartE2EDuration="3.866172318s" podCreationTimestamp="2025-12-07 09:26:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:23.843172804 +0000 UTC m=+1200.550491861" watchObservedRunningTime="2025-12-07 09:26:23.866172318 +0000 UTC m=+1200.573491335" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.888670 4838 scope.go:117] "RemoveContainer" containerID="ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9" Dec 07 09:26:23 crc kubenswrapper[4838]: E1207 09:26:23.889534 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9\": container with ID starting with ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9 not found: ID does not exist" containerID="ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.889577 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9"} err="failed to get container status \"ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9\": rpc error: code = NotFound desc = could not find container \"ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9\": container with ID starting with ba808f04b2979aa1ee3d0f49bb5b436292f86a2905233ccb6a1c21fcc619a2d9 not found: ID does not exist" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.889600 4838 scope.go:117] "RemoveContainer" containerID="ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a" Dec 07 09:26:23 crc kubenswrapper[4838]: E1207 09:26:23.892914 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a\": container with ID starting with ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a not found: ID does not exist" containerID="ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a" Dec 07 09:26:23 crc kubenswrapper[4838]: I1207 09:26:23.892948 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a"} err="failed to get container status \"ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a\": rpc error: code = NotFound desc = could not find container \"ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a\": container with ID starting with ccedfb3c872fae2cd5aee4be70ecceae450ef816b1bc519a9780fbd6b8474e0a not found: ID does not exist" Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.236108 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc\") pod \"1b96570d-092b-4374-9e9c-80c459790207\" (UID: \"1b96570d-092b-4374-9e9c-80c459790207\") " Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.237002 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b96570d-092b-4374-9e9c-80c459790207" (UID: "1b96570d-092b-4374-9e9c-80c459790207"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.339089 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b96570d-092b-4374-9e9c-80c459790207-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.398612 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578f9b9457-kfn5d"] Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.406061 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-578f9b9457-kfn5d"] Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.785091 4838 generic.go:334] "Generic (PLEG): container finished" podID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerID="f7eea783ea7f3e5918b2b537f29718db4c868b2fe5999776c60bd87363840a94" exitCode=0 Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.786104 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6242b9-8e57-4a38-b450-6ad5f9b804d6","Type":"ContainerDied","Data":"f7eea783ea7f3e5918b2b537f29718db4c868b2fe5999776c60bd87363840a94"} Dec 07 09:26:24 crc kubenswrapper[4838]: I1207 09:26:24.786138 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.182345 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.333761 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.630668 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b96570d-092b-4374-9e9c-80c459790207" path="/var/lib/kubelet/pods/1b96570d-092b-4374-9e9c-80c459790207/volumes" Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.765225 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7998cf6c8f-xbbtt" Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.824392 4838 generic.go:334] "Generic (PLEG): container finished" podID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerID="7c6a0f2185ddf0db2798a204c29a726b10a7bafe3f84d295229ae907feb14188" exitCode=0 Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.833080 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6242b9-8e57-4a38-b450-6ad5f9b804d6","Type":"ContainerDied","Data":"7c6a0f2185ddf0db2798a204c29a726b10a7bafe3f84d295229ae907feb14188"} Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.849521 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78dc87b64b-wtzk9"] Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.866120 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78dc87b64b-wtzk9" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-api" containerID="cri-o://9294585ebc68e5f3b84b5ab283abca2158c22fd3c3f8757861adc83256bacccf" gracePeriod=30 Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.866575 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78dc87b64b-wtzk9" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-httpd" containerID="cri-o://d4cb4afa0c8e6650789c385c48bad71ba1e1992d1a95196c9812a375dccd5cee" gracePeriod=30 Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.897655 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.980770 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data-custom\") pod \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.980934 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-combined-ca-bundle\") pod \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.980995 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-scripts\") pod \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.981093 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2qvm\" (UniqueName: \"kubernetes.io/projected/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-kube-api-access-n2qvm\") pod \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.981159 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data\") pod \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.981224 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-etc-machine-id\") pod \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\" (UID: \"9a6242b9-8e57-4a38-b450-6ad5f9b804d6\") " Dec 07 09:26:25 crc kubenswrapper[4838]: I1207 09:26:25.981664 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9a6242b9-8e57-4a38-b450-6ad5f9b804d6" (UID: "9a6242b9-8e57-4a38-b450-6ad5f9b804d6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.019479 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9a6242b9-8e57-4a38-b450-6ad5f9b804d6" (UID: "9a6242b9-8e57-4a38-b450-6ad5f9b804d6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.019731 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-scripts" (OuterVolumeSpecName: "scripts") pod "9a6242b9-8e57-4a38-b450-6ad5f9b804d6" (UID: "9a6242b9-8e57-4a38-b450-6ad5f9b804d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.033037 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-kube-api-access-n2qvm" (OuterVolumeSpecName: "kube-api-access-n2qvm") pod "9a6242b9-8e57-4a38-b450-6ad5f9b804d6" (UID: "9a6242b9-8e57-4a38-b450-6ad5f9b804d6"). InnerVolumeSpecName "kube-api-access-n2qvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.083175 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.083209 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2qvm\" (UniqueName: \"kubernetes.io/projected/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-kube-api-access-n2qvm\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.083237 4838 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.083246 4838 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.091259 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a6242b9-8e57-4a38-b450-6ad5f9b804d6" (UID: "9a6242b9-8e57-4a38-b450-6ad5f9b804d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.187178 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.234988 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data" (OuterVolumeSpecName: "config-data") pod "9a6242b9-8e57-4a38-b450-6ad5f9b804d6" (UID: "9a6242b9-8e57-4a38-b450-6ad5f9b804d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.288689 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6242b9-8e57-4a38-b450-6ad5f9b804d6-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.576355 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.612446 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9b574bc66-nsc2v" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.851647 4838 generic.go:334] "Generic (PLEG): container finished" podID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerID="d4cb4afa0c8e6650789c385c48bad71ba1e1992d1a95196c9812a375dccd5cee" exitCode=0 Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.851715 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc87b64b-wtzk9" event={"ID":"4224e6bd-236c-43d1-a0a1-55709ff562c7","Type":"ContainerDied","Data":"d4cb4afa0c8e6650789c385c48bad71ba1e1992d1a95196c9812a375dccd5cee"} Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.856539 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.856876 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6242b9-8e57-4a38-b450-6ad5f9b804d6","Type":"ContainerDied","Data":"7066c106e80818165324339f52ec22117e2e03ef3909605b48f1e1420ea65bcf"} Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.856965 4838 scope.go:117] "RemoveContainer" containerID="f7eea783ea7f3e5918b2b537f29718db4c868b2fe5999776c60bd87363840a94" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.905014 4838 scope.go:117] "RemoveContainer" containerID="7c6a0f2185ddf0db2798a204c29a726b10a7bafe3f84d295229ae907feb14188" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.915016 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.936351 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.944933 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:26 crc kubenswrapper[4838]: E1207 09:26:26.945257 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b96570d-092b-4374-9e9c-80c459790207" containerName="init" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.945272 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b96570d-092b-4374-9e9c-80c459790207" containerName="init" Dec 07 09:26:26 crc kubenswrapper[4838]: E1207 09:26:26.945291 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="probe" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.945297 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="probe" Dec 07 09:26:26 crc kubenswrapper[4838]: E1207 09:26:26.945322 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b96570d-092b-4374-9e9c-80c459790207" containerName="dnsmasq-dns" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.945328 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b96570d-092b-4374-9e9c-80c459790207" containerName="dnsmasq-dns" Dec 07 09:26:26 crc kubenswrapper[4838]: E1207 09:26:26.945337 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="cinder-scheduler" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.945343 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="cinder-scheduler" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.945493 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="cinder-scheduler" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.945507 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b96570d-092b-4374-9e9c-80c459790207" containerName="dnsmasq-dns" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.945518 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" containerName="probe" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.946339 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.951919 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 07 09:26:26 crc kubenswrapper[4838]: I1207 09:26:26.969927 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.018685 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.019112 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-config-data\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.019132 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.019210 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k8bl\" (UniqueName: \"kubernetes.io/projected/51d41156-ff39-4850-a562-5ba714cbd169-kube-api-access-7k8bl\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.019232 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-scripts\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.019284 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51d41156-ff39-4850-a562-5ba714cbd169-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.120345 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-config-data\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.120387 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.120455 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k8bl\" (UniqueName: \"kubernetes.io/projected/51d41156-ff39-4850-a562-5ba714cbd169-kube-api-access-7k8bl\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.120475 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-scripts\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.120526 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51d41156-ff39-4850-a562-5ba714cbd169-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.120562 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.120864 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51d41156-ff39-4850-a562-5ba714cbd169-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.124511 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.125238 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-scripts\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.125895 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.126597 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d41156-ff39-4850-a562-5ba714cbd169-config-data\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.138494 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k8bl\" (UniqueName: \"kubernetes.io/projected/51d41156-ff39-4850-a562-5ba714cbd169-kube-api-access-7k8bl\") pod \"cinder-scheduler-0\" (UID: \"51d41156-ff39-4850-a562-5ba714cbd169\") " pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.267730 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.626805 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a6242b9-8e57-4a38-b450-6ad5f9b804d6" path="/var/lib/kubelet/pods/9a6242b9-8e57-4a38-b450-6ad5f9b804d6/volumes" Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.843348 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 09:26:27 crc kubenswrapper[4838]: I1207 09:26:27.865280 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51d41156-ff39-4850-a562-5ba714cbd169","Type":"ContainerStarted","Data":"d4d088d6c76b5192b9149a594846a4b6b232dfb4358fb7b9a9e7c69d030fbc13"} Dec 07 09:26:28 crc kubenswrapper[4838]: I1207 09:26:28.898561 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51d41156-ff39-4850-a562-5ba714cbd169","Type":"ContainerStarted","Data":"fe5f853d68849197e49aa98fbdf6a0d00420ed42ddcbf884da9f2332f910f4bd"} Dec 07 09:26:29 crc kubenswrapper[4838]: I1207 09:26:29.929725 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51d41156-ff39-4850-a562-5ba714cbd169","Type":"ContainerStarted","Data":"a9d79a2f5231af9338127e6669344010eca62b34909767d773bbacaf7425d5cd"} Dec 07 09:26:29 crc kubenswrapper[4838]: I1207 09:26:29.940985 4838 generic.go:334] "Generic (PLEG): container finished" podID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerID="9294585ebc68e5f3b84b5ab283abca2158c22fd3c3f8757861adc83256bacccf" exitCode=0 Dec 07 09:26:29 crc kubenswrapper[4838]: I1207 09:26:29.941045 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc87b64b-wtzk9" event={"ID":"4224e6bd-236c-43d1-a0a1-55709ff562c7","Type":"ContainerDied","Data":"9294585ebc68e5f3b84b5ab283abca2158c22fd3c3f8757861adc83256bacccf"} Dec 07 09:26:29 crc kubenswrapper[4838]: I1207 09:26:29.962867 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.962798082 podStartE2EDuration="3.962798082s" podCreationTimestamp="2025-12-07 09:26:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:29.951402667 +0000 UTC m=+1206.658721685" watchObservedRunningTime="2025-12-07 09:26:29.962798082 +0000 UTC m=+1206.670117099" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.006783 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.075913 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-config\") pod \"4224e6bd-236c-43d1-a0a1-55709ff562c7\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.075972 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-ovndb-tls-certs\") pod \"4224e6bd-236c-43d1-a0a1-55709ff562c7\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.076007 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-httpd-config\") pod \"4224e6bd-236c-43d1-a0a1-55709ff562c7\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.076037 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s46d\" (UniqueName: \"kubernetes.io/projected/4224e6bd-236c-43d1-a0a1-55709ff562c7-kube-api-access-4s46d\") pod \"4224e6bd-236c-43d1-a0a1-55709ff562c7\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.076104 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-combined-ca-bundle\") pod \"4224e6bd-236c-43d1-a0a1-55709ff562c7\" (UID: \"4224e6bd-236c-43d1-a0a1-55709ff562c7\") " Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.112072 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4224e6bd-236c-43d1-a0a1-55709ff562c7" (UID: "4224e6bd-236c-43d1-a0a1-55709ff562c7"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.116964 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4224e6bd-236c-43d1-a0a1-55709ff562c7-kube-api-access-4s46d" (OuterVolumeSpecName: "kube-api-access-4s46d") pod "4224e6bd-236c-43d1-a0a1-55709ff562c7" (UID: "4224e6bd-236c-43d1-a0a1-55709ff562c7"). InnerVolumeSpecName "kube-api-access-4s46d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.182005 4838 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.182032 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s46d\" (UniqueName: \"kubernetes.io/projected/4224e6bd-236c-43d1-a0a1-55709ff562c7-kube-api-access-4s46d\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.196998 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4224e6bd-236c-43d1-a0a1-55709ff562c7" (UID: "4224e6bd-236c-43d1-a0a1-55709ff562c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.249657 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-config" (OuterVolumeSpecName: "config") pod "4224e6bd-236c-43d1-a0a1-55709ff562c7" (UID: "4224e6bd-236c-43d1-a0a1-55709ff562c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.257078 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4224e6bd-236c-43d1-a0a1-55709ff562c7" (UID: "4224e6bd-236c-43d1-a0a1-55709ff562c7"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.283703 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.283749 4838 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.283761 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4224e6bd-236c-43d1-a0a1-55709ff562c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.421339 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-654fd9b8c-czchm" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.951397 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc87b64b-wtzk9" event={"ID":"4224e6bd-236c-43d1-a0a1-55709ff562c7","Type":"ContainerDied","Data":"cdc146184d466ebff0ebd45fab22c01539b26694b666adbb0acb76c0c3a128aa"} Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.951472 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc87b64b-wtzk9" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.951710 4838 scope.go:117] "RemoveContainer" containerID="d4cb4afa0c8e6650789c385c48bad71ba1e1992d1a95196c9812a375dccd5cee" Dec 07 09:26:30 crc kubenswrapper[4838]: I1207 09:26:30.988373 4838 scope.go:117] "RemoveContainer" containerID="9294585ebc68e5f3b84b5ab283abca2158c22fd3c3f8757861adc83256bacccf" Dec 07 09:26:31 crc kubenswrapper[4838]: I1207 09:26:31.036996 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78dc87b64b-wtzk9"] Dec 07 09:26:31 crc kubenswrapper[4838]: I1207 09:26:31.052926 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-78dc87b64b-wtzk9"] Dec 07 09:26:31 crc kubenswrapper[4838]: I1207 09:26:31.624016 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" path="/var/lib/kubelet/pods/4224e6bd-236c-43d1-a0a1-55709ff562c7/volumes" Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.031825 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.268221 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.493163 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-59c4fb88bd-m7xbq" Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.551504 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-75749567f4-gxjqb"] Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.551731 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-75749567f4-gxjqb" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api-log" containerID="cri-o://8d1ab074842a6925aa4eb04268ce47241560a46a1d2c61e0dbb58fe003089638" gracePeriod=30 Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.552108 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-75749567f4-gxjqb" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api" containerID="cri-o://439ee46fbe938dcd548f618d4c5c8696442fd7758e234610b8685441bd07429d" gracePeriod=30 Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.974931 4838 generic.go:334] "Generic (PLEG): container finished" podID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerID="8d1ab074842a6925aa4eb04268ce47241560a46a1d2c61e0dbb58fe003089638" exitCode=143 Dec 07 09:26:32 crc kubenswrapper[4838]: I1207 09:26:32.976432 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75749567f4-gxjqb" event={"ID":"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c","Type":"ContainerDied","Data":"8d1ab074842a6925aa4eb04268ce47241560a46a1d2c61e0dbb58fe003089638"} Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.706780 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 07 09:26:33 crc kubenswrapper[4838]: E1207 09:26:33.707116 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-httpd" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.707129 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-httpd" Dec 07 09:26:33 crc kubenswrapper[4838]: E1207 09:26:33.707160 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-api" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.707168 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-api" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.707326 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-httpd" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.707340 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="4224e6bd-236c-43d1-a0a1-55709ff562c7" containerName="neutron-api" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.707957 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.710840 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.710844 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-fpqjm" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.712008 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.719911 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.783826 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/096a9643-43b6-4f09-8be1-cd739058c12e-openstack-config-secret\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.783869 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh4gc\" (UniqueName: \"kubernetes.io/projected/096a9643-43b6-4f09-8be1-cd739058c12e-kube-api-access-xh4gc\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.783967 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/096a9643-43b6-4f09-8be1-cd739058c12e-openstack-config\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.784020 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096a9643-43b6-4f09-8be1-cd739058c12e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.886112 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/096a9643-43b6-4f09-8be1-cd739058c12e-openstack-config\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.886204 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096a9643-43b6-4f09-8be1-cd739058c12e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.886300 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/096a9643-43b6-4f09-8be1-cd739058c12e-openstack-config-secret\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.886328 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh4gc\" (UniqueName: \"kubernetes.io/projected/096a9643-43b6-4f09-8be1-cd739058c12e-kube-api-access-xh4gc\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.887675 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/096a9643-43b6-4f09-8be1-cd739058c12e-openstack-config\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.892808 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096a9643-43b6-4f09-8be1-cd739058c12e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.893315 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/096a9643-43b6-4f09-8be1-cd739058c12e-openstack-config-secret\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:33 crc kubenswrapper[4838]: I1207 09:26:33.905344 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh4gc\" (UniqueName: \"kubernetes.io/projected/096a9643-43b6-4f09-8be1-cd739058c12e-kube-api-access-xh4gc\") pod \"openstackclient\" (UID: \"096a9643-43b6-4f09-8be1-cd739058c12e\") " pod="openstack/openstackclient" Dec 07 09:26:34 crc kubenswrapper[4838]: I1207 09:26:34.023982 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 07 09:26:34 crc kubenswrapper[4838]: I1207 09:26:34.129635 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 07 09:26:34 crc kubenswrapper[4838]: I1207 09:26:34.646918 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 07 09:26:34 crc kubenswrapper[4838]: I1207 09:26:34.992316 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"096a9643-43b6-4f09-8be1-cd739058c12e","Type":"ContainerStarted","Data":"ece77533bbc994f9ffa8e265b2e7551c9b2d0dbf8c57a0baa401e9a70eadb0e1"} Dec 07 09:26:35 crc kubenswrapper[4838]: I1207 09:26:35.758494 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-75749567f4-gxjqb" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:50232->10.217.0.146:9311: read: connection reset by peer" Dec 07 09:26:35 crc kubenswrapper[4838]: I1207 09:26:35.758524 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-75749567f4-gxjqb" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:50240->10.217.0.146:9311: read: connection reset by peer" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.006257 4838 generic.go:334] "Generic (PLEG): container finished" podID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerID="439ee46fbe938dcd548f618d4c5c8696442fd7758e234610b8685441bd07429d" exitCode=0 Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.006292 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75749567f4-gxjqb" event={"ID":"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c","Type":"ContainerDied","Data":"439ee46fbe938dcd548f618d4c5c8696442fd7758e234610b8685441bd07429d"} Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.125146 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.233015 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data\") pod \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.233087 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-combined-ca-bundle\") pod \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.233189 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data-custom\") pod \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.233223 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-logs\") pod \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.233301 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l98wz\" (UniqueName: \"kubernetes.io/projected/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-kube-api-access-l98wz\") pod \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\" (UID: \"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c\") " Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.235362 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-logs" (OuterVolumeSpecName: "logs") pod "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" (UID: "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.245068 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" (UID: "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.245371 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-kube-api-access-l98wz" (OuterVolumeSpecName: "kube-api-access-l98wz") pod "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" (UID: "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c"). InnerVolumeSpecName "kube-api-access-l98wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.269622 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" (UID: "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.288490 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data" (OuterVolumeSpecName: "config-data") pod "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" (UID: "9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.336605 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l98wz\" (UniqueName: \"kubernetes.io/projected/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-kube-api-access-l98wz\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.336638 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.336647 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.336656 4838 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:36 crc kubenswrapper[4838]: I1207 09:26:36.336664 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.019936 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75749567f4-gxjqb" event={"ID":"9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c","Type":"ContainerDied","Data":"59d17b2a422a362bc2185831e4d36402da217464d2af1b76d0d81a805b21ccf6"} Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.019992 4838 scope.go:117] "RemoveContainer" containerID="439ee46fbe938dcd548f618d4c5c8696442fd7758e234610b8685441bd07429d" Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.020129 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75749567f4-gxjqb" Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.057428 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-75749567f4-gxjqb"] Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.064297 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-75749567f4-gxjqb"] Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.067409 4838 scope.go:117] "RemoveContainer" containerID="8d1ab074842a6925aa4eb04268ce47241560a46a1d2c61e0dbb58fe003089638" Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.517732 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 07 09:26:37 crc kubenswrapper[4838]: I1207 09:26:37.625900 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" path="/var/lib/kubelet/pods/9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c/volumes" Dec 07 09:26:47 crc kubenswrapper[4838]: I1207 09:26:47.121571 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"096a9643-43b6-4f09-8be1-cd739058c12e","Type":"ContainerStarted","Data":"0d36ec6e4e403fe0eb97b9a9b2027f2fc1f046b8fe90322f4ebde370e9314d66"} Dec 07 09:26:47 crc kubenswrapper[4838]: I1207 09:26:47.145568 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.6683080869999998 podStartE2EDuration="14.145551868s" podCreationTimestamp="2025-12-07 09:26:33 +0000 UTC" firstStartedPulling="2025-12-07 09:26:34.64472013 +0000 UTC m=+1211.352039147" lastFinishedPulling="2025-12-07 09:26:46.121963901 +0000 UTC m=+1222.829282928" observedRunningTime="2025-12-07 09:26:47.14090205 +0000 UTC m=+1223.848221067" watchObservedRunningTime="2025-12-07 09:26:47.145551868 +0000 UTC m=+1223.852870885" Dec 07 09:26:47 crc kubenswrapper[4838]: I1207 09:26:47.153491 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 07 09:26:47 crc kubenswrapper[4838]: I1207 09:26:47.664458 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:48 crc kubenswrapper[4838]: I1207 09:26:48.128849 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-central-agent" containerID="cri-o://b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e" gracePeriod=30 Dec 07 09:26:48 crc kubenswrapper[4838]: I1207 09:26:48.128918 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="sg-core" containerID="cri-o://f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737" gracePeriod=30 Dec 07 09:26:48 crc kubenswrapper[4838]: I1207 09:26:48.128919 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="proxy-httpd" containerID="cri-o://341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145" gracePeriod=30 Dec 07 09:26:48 crc kubenswrapper[4838]: I1207 09:26:48.128967 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-notification-agent" containerID="cri-o://553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f" gracePeriod=30 Dec 07 09:26:49 crc kubenswrapper[4838]: I1207 09:26:49.137911 4838 generic.go:334] "Generic (PLEG): container finished" podID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerID="341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145" exitCode=0 Dec 07 09:26:49 crc kubenswrapper[4838]: I1207 09:26:49.138221 4838 generic.go:334] "Generic (PLEG): container finished" podID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerID="f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737" exitCode=2 Dec 07 09:26:49 crc kubenswrapper[4838]: I1207 09:26:49.138235 4838 generic.go:334] "Generic (PLEG): container finished" podID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerID="b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e" exitCode=0 Dec 07 09:26:49 crc kubenswrapper[4838]: I1207 09:26:49.137994 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerDied","Data":"341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145"} Dec 07 09:26:49 crc kubenswrapper[4838]: I1207 09:26:49.138304 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerDied","Data":"f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737"} Dec 07 09:26:49 crc kubenswrapper[4838]: I1207 09:26:49.138319 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerDied","Data":"b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e"} Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.744609 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.900392 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-combined-ca-bundle\") pod \"a0382eb7-0fc9-44ff-a191-3b9b15526127\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.900495 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-sg-core-conf-yaml\") pod \"a0382eb7-0fc9-44ff-a191-3b9b15526127\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.900564 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-scripts\") pod \"a0382eb7-0fc9-44ff-a191-3b9b15526127\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.900622 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-log-httpd\") pod \"a0382eb7-0fc9-44ff-a191-3b9b15526127\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.900677 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-config-data\") pod \"a0382eb7-0fc9-44ff-a191-3b9b15526127\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.900714 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gmxj\" (UniqueName: \"kubernetes.io/projected/a0382eb7-0fc9-44ff-a191-3b9b15526127-kube-api-access-2gmxj\") pod \"a0382eb7-0fc9-44ff-a191-3b9b15526127\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.900734 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-run-httpd\") pod \"a0382eb7-0fc9-44ff-a191-3b9b15526127\" (UID: \"a0382eb7-0fc9-44ff-a191-3b9b15526127\") " Dec 07 09:26:50 crc kubenswrapper[4838]: I1207 09:26:50.901530 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a0382eb7-0fc9-44ff-a191-3b9b15526127" (UID: "a0382eb7-0fc9-44ff-a191-3b9b15526127"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.306149 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a0382eb7-0fc9-44ff-a191-3b9b15526127" (UID: "a0382eb7-0fc9-44ff-a191-3b9b15526127"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.311579 4838 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.315312 4838 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0382eb7-0fc9-44ff-a191-3b9b15526127-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.312693 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-scripts" (OuterVolumeSpecName: "scripts") pod "a0382eb7-0fc9-44ff-a191-3b9b15526127" (UID: "a0382eb7-0fc9-44ff-a191-3b9b15526127"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.335630 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0382eb7-0fc9-44ff-a191-3b9b15526127-kube-api-access-2gmxj" (OuterVolumeSpecName: "kube-api-access-2gmxj") pod "a0382eb7-0fc9-44ff-a191-3b9b15526127" (UID: "a0382eb7-0fc9-44ff-a191-3b9b15526127"). InnerVolumeSpecName "kube-api-access-2gmxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.377424 4838 generic.go:334] "Generic (PLEG): container finished" podID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerID="553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f" exitCode=0 Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.377480 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerDied","Data":"553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f"} Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.377513 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0382eb7-0fc9-44ff-a191-3b9b15526127","Type":"ContainerDied","Data":"109b28ee2c45ba145bd9a71926e1e98c5143464e4594ef2a70506dce697dc427"} Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.377534 4838 scope.go:117] "RemoveContainer" containerID="341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.377688 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.380307 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a0382eb7-0fc9-44ff-a191-3b9b15526127" (UID: "a0382eb7-0fc9-44ff-a191-3b9b15526127"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.416728 4838 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.416770 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.416782 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gmxj\" (UniqueName: \"kubernetes.io/projected/a0382eb7-0fc9-44ff-a191-3b9b15526127-kube-api-access-2gmxj\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.497874 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0382eb7-0fc9-44ff-a191-3b9b15526127" (UID: "a0382eb7-0fc9-44ff-a191-3b9b15526127"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.516331 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-config-data" (OuterVolumeSpecName: "config-data") pod "a0382eb7-0fc9-44ff-a191-3b9b15526127" (UID: "a0382eb7-0fc9-44ff-a191-3b9b15526127"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.517639 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.517662 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0382eb7-0fc9-44ff-a191-3b9b15526127-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.589876 4838 scope.go:117] "RemoveContainer" containerID="f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.608573 4838 scope.go:117] "RemoveContainer" containerID="553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.632515 4838 scope.go:117] "RemoveContainer" containerID="b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.651224 4838 scope.go:117] "RemoveContainer" containerID="341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.651587 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145\": container with ID starting with 341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145 not found: ID does not exist" containerID="341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.651618 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145"} err="failed to get container status \"341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145\": rpc error: code = NotFound desc = could not find container \"341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145\": container with ID starting with 341cdb74336e3ea2222bd45647421138a780ea2985d5fa05918599497ffc7145 not found: ID does not exist" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.651636 4838 scope.go:117] "RemoveContainer" containerID="f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.651936 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737\": container with ID starting with f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737 not found: ID does not exist" containerID="f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.651955 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737"} err="failed to get container status \"f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737\": rpc error: code = NotFound desc = could not find container \"f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737\": container with ID starting with f42fe77a4b756ed4b9def3a115268b726fde45cb94636f81d12c88ebc935d737 not found: ID does not exist" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.651967 4838 scope.go:117] "RemoveContainer" containerID="553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.652175 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f\": container with ID starting with 553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f not found: ID does not exist" containerID="553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.652193 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f"} err="failed to get container status \"553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f\": rpc error: code = NotFound desc = could not find container \"553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f\": container with ID starting with 553d1527832914669a894194f9a6615de787063a9c8765c7ab30160f6001f32f not found: ID does not exist" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.652205 4838 scope.go:117] "RemoveContainer" containerID="b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.652410 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e\": container with ID starting with b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e not found: ID does not exist" containerID="b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.652429 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e"} err="failed to get container status \"b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e\": rpc error: code = NotFound desc = could not find container \"b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e\": container with ID starting with b31fe31a7b27e186429960052e0a2d5877838002850e15582d376db0e322168e not found: ID does not exist" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.710409 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.718308 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.742593 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.742939 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-notification-agent" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.742953 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-notification-agent" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.742965 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="proxy-httpd" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.742973 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="proxy-httpd" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.742984 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-central-agent" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.742990 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-central-agent" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.743013 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="sg-core" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743020 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="sg-core" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.743035 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api-log" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743042 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api-log" Dec 07 09:26:51 crc kubenswrapper[4838]: E1207 09:26:51.743062 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743068 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743212 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743223 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-notification-agent" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743233 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="ceilometer-central-agent" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743243 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fa1b1e7-72d2-4c4c-bb21-ef5c840a7c5c" containerName="barbican-api-log" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743256 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="proxy-httpd" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.743270 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" containerName="sg-core" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.744688 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.749689 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.749959 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.759121 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.759332 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="be998839-55b0-48de-9903-197e3684a4cf" containerName="kube-state-metrics" containerID="cri-o://9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef" gracePeriod=30 Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.765677 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.821558 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.821674 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-config-data\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.821722 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.821803 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-log-httpd\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.821903 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfc6b\" (UniqueName: \"kubernetes.io/projected/2270d2fe-8e5e-453a-83e1-371306ff8ef8-kube-api-access-tfc6b\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.821955 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-run-httpd\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.821990 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-scripts\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.923173 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfc6b\" (UniqueName: \"kubernetes.io/projected/2270d2fe-8e5e-453a-83e1-371306ff8ef8-kube-api-access-tfc6b\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.923224 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-run-httpd\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.923256 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-scripts\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.923286 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.923341 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-config-data\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.923376 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.923417 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-log-httpd\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.924445 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-run-httpd\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.924575 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-log-httpd\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.934859 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-config-data\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.936040 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.936272 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.936550 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-scripts\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:51 crc kubenswrapper[4838]: I1207 09:26:51.940197 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfc6b\" (UniqueName: \"kubernetes.io/projected/2270d2fe-8e5e-453a-83e1-371306ff8ef8-kube-api-access-tfc6b\") pod \"ceilometer-0\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " pod="openstack/ceilometer-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.104064 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.204554 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.241102 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwn9b\" (UniqueName: \"kubernetes.io/projected/be998839-55b0-48de-9903-197e3684a4cf-kube-api-access-mwn9b\") pod \"be998839-55b0-48de-9903-197e3684a4cf\" (UID: \"be998839-55b0-48de-9903-197e3684a4cf\") " Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.261178 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be998839-55b0-48de-9903-197e3684a4cf-kube-api-access-mwn9b" (OuterVolumeSpecName: "kube-api-access-mwn9b") pod "be998839-55b0-48de-9903-197e3684a4cf" (UID: "be998839-55b0-48de-9903-197e3684a4cf"). InnerVolumeSpecName "kube-api-access-mwn9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.342866 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwn9b\" (UniqueName: \"kubernetes.io/projected/be998839-55b0-48de-9903-197e3684a4cf-kube-api-access-mwn9b\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.412701 4838 generic.go:334] "Generic (PLEG): container finished" podID="be998839-55b0-48de-9903-197e3684a4cf" containerID="9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef" exitCode=2 Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.412757 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be998839-55b0-48de-9903-197e3684a4cf","Type":"ContainerDied","Data":"9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef"} Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.412784 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be998839-55b0-48de-9903-197e3684a4cf","Type":"ContainerDied","Data":"9d104a1501adf71063cf8a0f02f1b02023df463c279bef243677e22b2075a0e1"} Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.412803 4838 scope.go:117] "RemoveContainer" containerID="9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.412978 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.483966 4838 scope.go:117] "RemoveContainer" containerID="9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef" Dec 07 09:26:52 crc kubenswrapper[4838]: E1207 09:26:52.484807 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef\": container with ID starting with 9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef not found: ID does not exist" containerID="9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.484850 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef"} err="failed to get container status \"9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef\": rpc error: code = NotFound desc = could not find container \"9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef\": container with ID starting with 9b02c5acc68954590b1e429b8ffcea2524141940d39a39bb91e6c926ce3aeeef not found: ID does not exist" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.486430 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.502028 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.510509 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:26:52 crc kubenswrapper[4838]: E1207 09:26:52.511261 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be998839-55b0-48de-9903-197e3684a4cf" containerName="kube-state-metrics" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.511318 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="be998839-55b0-48de-9903-197e3684a4cf" containerName="kube-state-metrics" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.511617 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="be998839-55b0-48de-9903-197e3684a4cf" containerName="kube-state-metrics" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.513149 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.517532 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.517832 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.519669 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.555520 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv9sj\" (UniqueName: \"kubernetes.io/projected/a10a8e53-204b-48d6-aaef-809d1847f251-kube-api-access-gv9sj\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.555914 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.556024 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.556205 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.639095 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.658769 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.658858 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv9sj\" (UniqueName: \"kubernetes.io/projected/a10a8e53-204b-48d6-aaef-809d1847f251-kube-api-access-gv9sj\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.658888 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.658933 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.667152 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.671327 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.673323 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10a8e53-204b-48d6-aaef-809d1847f251-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.681456 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv9sj\" (UniqueName: \"kubernetes.io/projected/a10a8e53-204b-48d6-aaef-809d1847f251-kube-api-access-gv9sj\") pod \"kube-state-metrics-0\" (UID: \"a10a8e53-204b-48d6-aaef-809d1847f251\") " pod="openstack/kube-state-metrics-0" Dec 07 09:26:52 crc kubenswrapper[4838]: I1207 09:26:52.833202 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 09:26:53 crc kubenswrapper[4838]: I1207 09:26:53.323155 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 09:26:53 crc kubenswrapper[4838]: W1207 09:26:53.324310 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda10a8e53_204b_48d6_aaef_809d1847f251.slice/crio-f750033f074c949c72ac77fdf587bc00dcda127fb091289668cab1d449dd1743 WatchSource:0}: Error finding container f750033f074c949c72ac77fdf587bc00dcda127fb091289668cab1d449dd1743: Status 404 returned error can't find the container with id f750033f074c949c72ac77fdf587bc00dcda127fb091289668cab1d449dd1743 Dec 07 09:26:53 crc kubenswrapper[4838]: I1207 09:26:53.420852 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a10a8e53-204b-48d6-aaef-809d1847f251","Type":"ContainerStarted","Data":"f750033f074c949c72ac77fdf587bc00dcda127fb091289668cab1d449dd1743"} Dec 07 09:26:53 crc kubenswrapper[4838]: I1207 09:26:53.422044 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerStarted","Data":"18c166d4230430f3ba50efbced836d2111f27b08c659b2fc80900562cdb58809"} Dec 07 09:26:53 crc kubenswrapper[4838]: I1207 09:26:53.625610 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0382eb7-0fc9-44ff-a191-3b9b15526127" path="/var/lib/kubelet/pods/a0382eb7-0fc9-44ff-a191-3b9b15526127/volumes" Dec 07 09:26:53 crc kubenswrapper[4838]: I1207 09:26:53.626596 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be998839-55b0-48de-9903-197e3684a4cf" path="/var/lib/kubelet/pods/be998839-55b0-48de-9903-197e3684a4cf/volumes" Dec 07 09:26:53 crc kubenswrapper[4838]: I1207 09:26:53.728676 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.432457 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerStarted","Data":"b4a6dd3a4191ba0addd0839e6003f513db1f7727a496c01ac32786e8b25662f6"} Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.433749 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a10a8e53-204b-48d6-aaef-809d1847f251","Type":"ContainerStarted","Data":"64db54f4724913566c956e878d7d3c8b444e6a60f362c39f7622ec98fd67651a"} Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.434353 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.438461 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-ztr5c"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.439560 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.452674 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ztr5c"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.484698 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.118279479 podStartE2EDuration="2.484667353s" podCreationTimestamp="2025-12-07 09:26:52 +0000 UTC" firstStartedPulling="2025-12-07 09:26:53.326927922 +0000 UTC m=+1230.034246939" lastFinishedPulling="2025-12-07 09:26:53.693315796 +0000 UTC m=+1230.400634813" observedRunningTime="2025-12-07 09:26:54.477786663 +0000 UTC m=+1231.185105680" watchObservedRunningTime="2025-12-07 09:26:54.484667353 +0000 UTC m=+1231.191986370" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.511156 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-operator-scripts\") pod \"nova-api-db-create-ztr5c\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.511298 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x957\" (UniqueName: \"kubernetes.io/projected/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-kube-api-access-2x957\") pod \"nova-api-db-create-ztr5c\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.538015 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-975b7"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.539123 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.559541 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-975b7"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.639563 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1f91-account-create-update-gqmgj"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.642995 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.645876 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.663673 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-operator-scripts\") pod \"nova-api-db-create-ztr5c\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.664075 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w4g6\" (UniqueName: \"kubernetes.io/projected/99e90708-7b16-4279-bb16-423c616b152c-kube-api-access-2w4g6\") pod \"nova-cell0-db-create-975b7\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.664931 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x957\" (UniqueName: \"kubernetes.io/projected/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-kube-api-access-2x957\") pod \"nova-api-db-create-ztr5c\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.665112 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99e90708-7b16-4279-bb16-423c616b152c-operator-scripts\") pod \"nova-cell0-db-create-975b7\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.667475 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-operator-scripts\") pod \"nova-api-db-create-ztr5c\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.696967 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x957\" (UniqueName: \"kubernetes.io/projected/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-kube-api-access-2x957\") pod \"nova-api-db-create-ztr5c\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.733757 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1f91-account-create-update-gqmgj"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.761659 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.767913 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-98sxs"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.769975 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-operator-scripts\") pod \"nova-api-1f91-account-create-update-gqmgj\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.770030 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ln4t\" (UniqueName: \"kubernetes.io/projected/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-kube-api-access-7ln4t\") pod \"nova-api-1f91-account-create-update-gqmgj\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.770140 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w4g6\" (UniqueName: \"kubernetes.io/projected/99e90708-7b16-4279-bb16-423c616b152c-kube-api-access-2w4g6\") pod \"nova-cell0-db-create-975b7\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.770213 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99e90708-7b16-4279-bb16-423c616b152c-operator-scripts\") pod \"nova-cell0-db-create-975b7\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.771224 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99e90708-7b16-4279-bb16-423c616b152c-operator-scripts\") pod \"nova-cell0-db-create-975b7\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.781380 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-98sxs"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.779608 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.821425 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w4g6\" (UniqueName: \"kubernetes.io/projected/99e90708-7b16-4279-bb16-423c616b152c-kube-api-access-2w4g6\") pod \"nova-cell0-db-create-975b7\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.822792 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ef37-account-create-update-djqnb"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.823915 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.828645 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.838939 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ef37-account-create-update-djqnb"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.862470 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.875727 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-operator-scripts\") pod \"nova-cell0-ef37-account-create-update-djqnb\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.875858 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-operator-scripts\") pod \"nova-api-1f91-account-create-update-gqmgj\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.875947 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ln4t\" (UniqueName: \"kubernetes.io/projected/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-kube-api-access-7ln4t\") pod \"nova-api-1f91-account-create-update-gqmgj\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.876039 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjsrm\" (UniqueName: \"kubernetes.io/projected/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-kube-api-access-vjsrm\") pod \"nova-cell0-ef37-account-create-update-djqnb\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.876147 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhnxr\" (UniqueName: \"kubernetes.io/projected/779f6638-c56b-462b-a574-21d3a2e08c28-kube-api-access-jhnxr\") pod \"nova-cell1-db-create-98sxs\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.876173 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/779f6638-c56b-462b-a574-21d3a2e08c28-operator-scripts\") pod \"nova-cell1-db-create-98sxs\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.881292 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-operator-scripts\") pod \"nova-api-1f91-account-create-update-gqmgj\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.906650 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ln4t\" (UniqueName: \"kubernetes.io/projected/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-kube-api-access-7ln4t\") pod \"nova-api-1f91-account-create-update-gqmgj\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.964943 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-c71d-account-create-update-dtbgj"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.966275 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.969177 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.974222 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c71d-account-create-update-dtbgj"] Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.977572 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhnxr\" (UniqueName: \"kubernetes.io/projected/779f6638-c56b-462b-a574-21d3a2e08c28-kube-api-access-jhnxr\") pod \"nova-cell1-db-create-98sxs\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.977618 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/779f6638-c56b-462b-a574-21d3a2e08c28-operator-scripts\") pod \"nova-cell1-db-create-98sxs\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.977672 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-operator-scripts\") pod \"nova-cell0-ef37-account-create-update-djqnb\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.977708 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd80aba1-e836-4c0f-8eba-2999e284db44-operator-scripts\") pod \"nova-cell1-c71d-account-create-update-dtbgj\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.977864 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brltq\" (UniqueName: \"kubernetes.io/projected/dd80aba1-e836-4c0f-8eba-2999e284db44-kube-api-access-brltq\") pod \"nova-cell1-c71d-account-create-update-dtbgj\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.978216 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjsrm\" (UniqueName: \"kubernetes.io/projected/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-kube-api-access-vjsrm\") pod \"nova-cell0-ef37-account-create-update-djqnb\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.978916 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-operator-scripts\") pod \"nova-cell0-ef37-account-create-update-djqnb\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.979206 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/779f6638-c56b-462b-a574-21d3a2e08c28-operator-scripts\") pod \"nova-cell1-db-create-98sxs\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:54 crc kubenswrapper[4838]: I1207 09:26:54.979335 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.005185 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhnxr\" (UniqueName: \"kubernetes.io/projected/779f6638-c56b-462b-a574-21d3a2e08c28-kube-api-access-jhnxr\") pod \"nova-cell1-db-create-98sxs\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.005199 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjsrm\" (UniqueName: \"kubernetes.io/projected/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-kube-api-access-vjsrm\") pod \"nova-cell0-ef37-account-create-update-djqnb\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.080980 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd80aba1-e836-4c0f-8eba-2999e284db44-operator-scripts\") pod \"nova-cell1-c71d-account-create-update-dtbgj\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.081339 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brltq\" (UniqueName: \"kubernetes.io/projected/dd80aba1-e836-4c0f-8eba-2999e284db44-kube-api-access-brltq\") pod \"nova-cell1-c71d-account-create-update-dtbgj\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.082248 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd80aba1-e836-4c0f-8eba-2999e284db44-operator-scripts\") pod \"nova-cell1-c71d-account-create-update-dtbgj\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.099724 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brltq\" (UniqueName: \"kubernetes.io/projected/dd80aba1-e836-4c0f-8eba-2999e284db44-kube-api-access-brltq\") pod \"nova-cell1-c71d-account-create-update-dtbgj\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.209086 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.230504 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.307037 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.418411 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ztr5c"] Dec 07 09:26:55 crc kubenswrapper[4838]: W1207 09:26:55.465219 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fe4a29d_3f3d_4cc3_884f_00f111d7d651.slice/crio-59b9732286caad043e9546db751b1be7e63f2f52d8a127322aa62a90548c68a4 WatchSource:0}: Error finding container 59b9732286caad043e9546db751b1be7e63f2f52d8a127322aa62a90548c68a4: Status 404 returned error can't find the container with id 59b9732286caad043e9546db751b1be7e63f2f52d8a127322aa62a90548c68a4 Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.479083 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerStarted","Data":"e6ed6db9299997cec854112736dad69b194ae457392c3fedb07261b3498ce8e4"} Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.603135 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-975b7"] Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.645417 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1f91-account-create-update-gqmgj"] Dec 07 09:26:55 crc kubenswrapper[4838]: W1207 09:26:55.700119 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90e334fc_b52e_4b90_bcd6_736bdc1fcd3a.slice/crio-2e64e8acbb8f4d2b2979535b220ba7768b0729b49faea92359b807fe70a05225 WatchSource:0}: Error finding container 2e64e8acbb8f4d2b2979535b220ba7768b0729b49faea92359b807fe70a05225: Status 404 returned error can't find the container with id 2e64e8acbb8f4d2b2979535b220ba7768b0729b49faea92359b807fe70a05225 Dec 07 09:26:55 crc kubenswrapper[4838]: I1207 09:26:55.888782 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ef37-account-create-update-djqnb"] Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.077847 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-98sxs"] Dec 07 09:26:56 crc kubenswrapper[4838]: W1207 09:26:56.088413 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod779f6638_c56b_462b_a574_21d3a2e08c28.slice/crio-35628422b129d7f8b72594a56b30a303533f9f693ba8ab39c3cbf450fa6cf809 WatchSource:0}: Error finding container 35628422b129d7f8b72594a56b30a303533f9f693ba8ab39c3cbf450fa6cf809: Status 404 returned error can't find the container with id 35628422b129d7f8b72594a56b30a303533f9f693ba8ab39c3cbf450fa6cf809 Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.093052 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c71d-account-create-update-dtbgj"] Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.498352 4838 generic.go:334] "Generic (PLEG): container finished" podID="7fe4a29d-3f3d-4cc3-884f-00f111d7d651" containerID="c7ec1855907e09fdb26e3e71384c14e52427ce6c0a56f97408e3fe2c5073fb7d" exitCode=0 Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.498672 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ztr5c" event={"ID":"7fe4a29d-3f3d-4cc3-884f-00f111d7d651","Type":"ContainerDied","Data":"c7ec1855907e09fdb26e3e71384c14e52427ce6c0a56f97408e3fe2c5073fb7d"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.498703 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ztr5c" event={"ID":"7fe4a29d-3f3d-4cc3-884f-00f111d7d651","Type":"ContainerStarted","Data":"59b9732286caad043e9546db751b1be7e63f2f52d8a127322aa62a90548c68a4"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.502546 4838 generic.go:334] "Generic (PLEG): container finished" podID="90e334fc-b52e-4b90-bcd6-736bdc1fcd3a" containerID="ea9efd8dc3bbbbbd175f07376d3f441b917ce8f4d6002a9071bf2234357e05d7" exitCode=0 Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.502613 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1f91-account-create-update-gqmgj" event={"ID":"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a","Type":"ContainerDied","Data":"ea9efd8dc3bbbbbd175f07376d3f441b917ce8f4d6002a9071bf2234357e05d7"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.502640 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1f91-account-create-update-gqmgj" event={"ID":"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a","Type":"ContainerStarted","Data":"2e64e8acbb8f4d2b2979535b220ba7768b0729b49faea92359b807fe70a05225"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.505406 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" event={"ID":"dd80aba1-e836-4c0f-8eba-2999e284db44","Type":"ContainerStarted","Data":"5579b735ad3e68795d3aaf69421e09ac15e0ff5575a5d4fc49b15be5f531e2ad"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.505436 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" event={"ID":"dd80aba1-e836-4c0f-8eba-2999e284db44","Type":"ContainerStarted","Data":"3599f1056c197920d4bf296a8f12aa1a83b418ce08cb03b56072d18328a27d4b"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.510741 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-98sxs" event={"ID":"779f6638-c56b-462b-a574-21d3a2e08c28","Type":"ContainerStarted","Data":"0763ab69bbf4838c83b1557c1944716eacb72716e08d734380faf656afae2b9e"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.510791 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-98sxs" event={"ID":"779f6638-c56b-462b-a574-21d3a2e08c28","Type":"ContainerStarted","Data":"35628422b129d7f8b72594a56b30a303533f9f693ba8ab39c3cbf450fa6cf809"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.516487 4838 generic.go:334] "Generic (PLEG): container finished" podID="99e90708-7b16-4279-bb16-423c616b152c" containerID="9f65f736b1524df7292ee3e57c12ec4628e3c67c76e5d01ab4788811b0e44b6f" exitCode=0 Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.516588 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-975b7" event={"ID":"99e90708-7b16-4279-bb16-423c616b152c","Type":"ContainerDied","Data":"9f65f736b1524df7292ee3e57c12ec4628e3c67c76e5d01ab4788811b0e44b6f"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.516648 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-975b7" event={"ID":"99e90708-7b16-4279-bb16-423c616b152c","Type":"ContainerStarted","Data":"2c72fb38bd432338b3b5fd66f3eea8d0bda06126121a1626a791226d60af6377"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.519186 4838 generic.go:334] "Generic (PLEG): container finished" podID="6c25b096-cc08-4a86-bc3c-b5c8601b3f71" containerID="7c5b1ddc9b76a5884073df34fc8a662a7c9011f864542fc08d94232f96a7b787" exitCode=0 Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.519269 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef37-account-create-update-djqnb" event={"ID":"6c25b096-cc08-4a86-bc3c-b5c8601b3f71","Type":"ContainerDied","Data":"7c5b1ddc9b76a5884073df34fc8a662a7c9011f864542fc08d94232f96a7b787"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.519322 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef37-account-create-update-djqnb" event={"ID":"6c25b096-cc08-4a86-bc3c-b5c8601b3f71","Type":"ContainerStarted","Data":"04bf17ecb309184fccb3671ad0a63042401d5dd6acabe23b546d22568f170afe"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.522374 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerStarted","Data":"a876029f334ad4701134b0979763a72f1b1ef0693032afa9fd785f733086db91"} Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.558116 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-98sxs" podStartSLOduration=2.558098544 podStartE2EDuration="2.558098544s" podCreationTimestamp="2025-12-07 09:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:56.548651686 +0000 UTC m=+1233.255970693" watchObservedRunningTime="2025-12-07 09:26:56.558098544 +0000 UTC m=+1233.265417561" Dec 07 09:26:56 crc kubenswrapper[4838]: I1207 09:26:56.588684 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" podStartSLOduration=2.588666568 podStartE2EDuration="2.588666568s" podCreationTimestamp="2025-12-07 09:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:26:56.563797344 +0000 UTC m=+1233.271116361" watchObservedRunningTime="2025-12-07 09:26:56.588666568 +0000 UTC m=+1233.295985585" Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.532601 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerStarted","Data":"a2cfd138493be75e4acd0d83a1a46029012557bc8df0db42b2f8667ea813b65c"} Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.532769 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-central-agent" containerID="cri-o://b4a6dd3a4191ba0addd0839e6003f513db1f7727a496c01ac32786e8b25662f6" gracePeriod=30 Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.532912 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-notification-agent" containerID="cri-o://e6ed6db9299997cec854112736dad69b194ae457392c3fedb07261b3498ce8e4" gracePeriod=30 Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.532920 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="sg-core" containerID="cri-o://a876029f334ad4701134b0979763a72f1b1ef0693032afa9fd785f733086db91" gracePeriod=30 Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.532908 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="proxy-httpd" containerID="cri-o://a2cfd138493be75e4acd0d83a1a46029012557bc8df0db42b2f8667ea813b65c" gracePeriod=30 Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.533433 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.536639 4838 generic.go:334] "Generic (PLEG): container finished" podID="dd80aba1-e836-4c0f-8eba-2999e284db44" containerID="5579b735ad3e68795d3aaf69421e09ac15e0ff5575a5d4fc49b15be5f531e2ad" exitCode=0 Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.536910 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" event={"ID":"dd80aba1-e836-4c0f-8eba-2999e284db44","Type":"ContainerDied","Data":"5579b735ad3e68795d3aaf69421e09ac15e0ff5575a5d4fc49b15be5f531e2ad"} Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.539843 4838 generic.go:334] "Generic (PLEG): container finished" podID="779f6638-c56b-462b-a574-21d3a2e08c28" containerID="0763ab69bbf4838c83b1557c1944716eacb72716e08d734380faf656afae2b9e" exitCode=0 Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.540011 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-98sxs" event={"ID":"779f6638-c56b-462b-a574-21d3a2e08c28","Type":"ContainerDied","Data":"0763ab69bbf4838c83b1557c1944716eacb72716e08d734380faf656afae2b9e"} Dec 07 09:26:57 crc kubenswrapper[4838]: I1207 09:26:57.559460 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.248330439 podStartE2EDuration="6.55943885s" podCreationTimestamp="2025-12-07 09:26:51 +0000 UTC" firstStartedPulling="2025-12-07 09:26:52.644969686 +0000 UTC m=+1229.352288703" lastFinishedPulling="2025-12-07 09:26:56.956078097 +0000 UTC m=+1233.663397114" observedRunningTime="2025-12-07 09:26:57.559018479 +0000 UTC m=+1234.266337486" watchObservedRunningTime="2025-12-07 09:26:57.55943885 +0000 UTC m=+1234.266757867" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.115599 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.137778 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.146850 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.149460 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.271349 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w4g6\" (UniqueName: \"kubernetes.io/projected/99e90708-7b16-4279-bb16-423c616b152c-kube-api-access-2w4g6\") pod \"99e90708-7b16-4279-bb16-423c616b152c\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.271411 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x957\" (UniqueName: \"kubernetes.io/projected/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-kube-api-access-2x957\") pod \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.271497 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-operator-scripts\") pod \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\" (UID: \"7fe4a29d-3f3d-4cc3-884f-00f111d7d651\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.271524 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-operator-scripts\") pod \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.271552 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjsrm\" (UniqueName: \"kubernetes.io/projected/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-kube-api-access-vjsrm\") pod \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\" (UID: \"6c25b096-cc08-4a86-bc3c-b5c8601b3f71\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.271661 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99e90708-7b16-4279-bb16-423c616b152c-operator-scripts\") pod \"99e90708-7b16-4279-bb16-423c616b152c\" (UID: \"99e90708-7b16-4279-bb16-423c616b152c\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.271748 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ln4t\" (UniqueName: \"kubernetes.io/projected/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-kube-api-access-7ln4t\") pod \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.272398 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7fe4a29d-3f3d-4cc3-884f-00f111d7d651" (UID: "7fe4a29d-3f3d-4cc3-884f-00f111d7d651"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.272495 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c25b096-cc08-4a86-bc3c-b5c8601b3f71" (UID: "6c25b096-cc08-4a86-bc3c-b5c8601b3f71"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.272573 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-operator-scripts\") pod \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\" (UID: \"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.273302 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99e90708-7b16-4279-bb16-423c616b152c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99e90708-7b16-4279-bb16-423c616b152c" (UID: "99e90708-7b16-4279-bb16-423c616b152c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.273620 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90e334fc-b52e-4b90-bcd6-736bdc1fcd3a" (UID: "90e334fc-b52e-4b90-bcd6-736bdc1fcd3a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.274953 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.274976 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.275007 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99e90708-7b16-4279-bb16-423c616b152c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.275018 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.280966 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e90708-7b16-4279-bb16-423c616b152c-kube-api-access-2w4g6" (OuterVolumeSpecName: "kube-api-access-2w4g6") pod "99e90708-7b16-4279-bb16-423c616b152c" (UID: "99e90708-7b16-4279-bb16-423c616b152c"). InnerVolumeSpecName "kube-api-access-2w4g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.281011 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-kube-api-access-2x957" (OuterVolumeSpecName: "kube-api-access-2x957") pod "7fe4a29d-3f3d-4cc3-884f-00f111d7d651" (UID: "7fe4a29d-3f3d-4cc3-884f-00f111d7d651"). InnerVolumeSpecName "kube-api-access-2x957". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.281081 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-kube-api-access-7ln4t" (OuterVolumeSpecName: "kube-api-access-7ln4t") pod "90e334fc-b52e-4b90-bcd6-736bdc1fcd3a" (UID: "90e334fc-b52e-4b90-bcd6-736bdc1fcd3a"). InnerVolumeSpecName "kube-api-access-7ln4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.281423 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-kube-api-access-vjsrm" (OuterVolumeSpecName: "kube-api-access-vjsrm") pod "6c25b096-cc08-4a86-bc3c-b5c8601b3f71" (UID: "6c25b096-cc08-4a86-bc3c-b5c8601b3f71"). InnerVolumeSpecName "kube-api-access-vjsrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.377155 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjsrm\" (UniqueName: \"kubernetes.io/projected/6c25b096-cc08-4a86-bc3c-b5c8601b3f71-kube-api-access-vjsrm\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.377224 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ln4t\" (UniqueName: \"kubernetes.io/projected/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a-kube-api-access-7ln4t\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.377237 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w4g6\" (UniqueName: \"kubernetes.io/projected/99e90708-7b16-4279-bb16-423c616b152c-kube-api-access-2w4g6\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.377248 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x957\" (UniqueName: \"kubernetes.io/projected/7fe4a29d-3f3d-4cc3-884f-00f111d7d651-kube-api-access-2x957\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.548407 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ztr5c" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.548420 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ztr5c" event={"ID":"7fe4a29d-3f3d-4cc3-884f-00f111d7d651","Type":"ContainerDied","Data":"59b9732286caad043e9546db751b1be7e63f2f52d8a127322aa62a90548c68a4"} Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.548484 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59b9732286caad043e9546db751b1be7e63f2f52d8a127322aa62a90548c68a4" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.550674 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1f91-account-create-update-gqmgj" event={"ID":"90e334fc-b52e-4b90-bcd6-736bdc1fcd3a","Type":"ContainerDied","Data":"2e64e8acbb8f4d2b2979535b220ba7768b0729b49faea92359b807fe70a05225"} Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.550700 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e64e8acbb8f4d2b2979535b220ba7768b0729b49faea92359b807fe70a05225" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.550762 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1f91-account-create-update-gqmgj" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.557585 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-975b7" event={"ID":"99e90708-7b16-4279-bb16-423c616b152c","Type":"ContainerDied","Data":"2c72fb38bd432338b3b5fd66f3eea8d0bda06126121a1626a791226d60af6377"} Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.557623 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c72fb38bd432338b3b5fd66f3eea8d0bda06126121a1626a791226d60af6377" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.557629 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-975b7" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.559387 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef37-account-create-update-djqnb" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.559384 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef37-account-create-update-djqnb" event={"ID":"6c25b096-cc08-4a86-bc3c-b5c8601b3f71","Type":"ContainerDied","Data":"04bf17ecb309184fccb3671ad0a63042401d5dd6acabe23b546d22568f170afe"} Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.559425 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04bf17ecb309184fccb3671ad0a63042401d5dd6acabe23b546d22568f170afe" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.562868 4838 generic.go:334] "Generic (PLEG): container finished" podID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerID="a2cfd138493be75e4acd0d83a1a46029012557bc8df0db42b2f8667ea813b65c" exitCode=0 Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.562888 4838 generic.go:334] "Generic (PLEG): container finished" podID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerID="a876029f334ad4701134b0979763a72f1b1ef0693032afa9fd785f733086db91" exitCode=2 Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.562896 4838 generic.go:334] "Generic (PLEG): container finished" podID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerID="e6ed6db9299997cec854112736dad69b194ae457392c3fedb07261b3498ce8e4" exitCode=0 Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.562996 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerDied","Data":"a2cfd138493be75e4acd0d83a1a46029012557bc8df0db42b2f8667ea813b65c"} Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.563033 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerDied","Data":"a876029f334ad4701134b0979763a72f1b1ef0693032afa9fd785f733086db91"} Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.563045 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerDied","Data":"e6ed6db9299997cec854112736dad69b194ae457392c3fedb07261b3498ce8e4"} Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.908573 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.997006 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.997542 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd80aba1-e836-4c0f-8eba-2999e284db44-operator-scripts\") pod \"dd80aba1-e836-4c0f-8eba-2999e284db44\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.997615 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brltq\" (UniqueName: \"kubernetes.io/projected/dd80aba1-e836-4c0f-8eba-2999e284db44-kube-api-access-brltq\") pod \"dd80aba1-e836-4c0f-8eba-2999e284db44\" (UID: \"dd80aba1-e836-4c0f-8eba-2999e284db44\") " Dec 07 09:26:58 crc kubenswrapper[4838]: I1207 09:26:58.998886 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd80aba1-e836-4c0f-8eba-2999e284db44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd80aba1-e836-4c0f-8eba-2999e284db44" (UID: "dd80aba1-e836-4c0f-8eba-2999e284db44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.004888 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd80aba1-e836-4c0f-8eba-2999e284db44-kube-api-access-brltq" (OuterVolumeSpecName: "kube-api-access-brltq") pod "dd80aba1-e836-4c0f-8eba-2999e284db44" (UID: "dd80aba1-e836-4c0f-8eba-2999e284db44"). InnerVolumeSpecName "kube-api-access-brltq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.099512 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/779f6638-c56b-462b-a574-21d3a2e08c28-operator-scripts\") pod \"779f6638-c56b-462b-a574-21d3a2e08c28\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.099641 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhnxr\" (UniqueName: \"kubernetes.io/projected/779f6638-c56b-462b-a574-21d3a2e08c28-kube-api-access-jhnxr\") pod \"779f6638-c56b-462b-a574-21d3a2e08c28\" (UID: \"779f6638-c56b-462b-a574-21d3a2e08c28\") " Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.099967 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/779f6638-c56b-462b-a574-21d3a2e08c28-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "779f6638-c56b-462b-a574-21d3a2e08c28" (UID: "779f6638-c56b-462b-a574-21d3a2e08c28"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.100170 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/779f6638-c56b-462b-a574-21d3a2e08c28-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.100188 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd80aba1-e836-4c0f-8eba-2999e284db44-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.100198 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brltq\" (UniqueName: \"kubernetes.io/projected/dd80aba1-e836-4c0f-8eba-2999e284db44-kube-api-access-brltq\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.104087 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/779f6638-c56b-462b-a574-21d3a2e08c28-kube-api-access-jhnxr" (OuterVolumeSpecName: "kube-api-access-jhnxr") pod "779f6638-c56b-462b-a574-21d3a2e08c28" (UID: "779f6638-c56b-462b-a574-21d3a2e08c28"). InnerVolumeSpecName "kube-api-access-jhnxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.201187 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhnxr\" (UniqueName: \"kubernetes.io/projected/779f6638-c56b-462b-a574-21d3a2e08c28-kube-api-access-jhnxr\") on node \"crc\" DevicePath \"\"" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.578994 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" event={"ID":"dd80aba1-e836-4c0f-8eba-2999e284db44","Type":"ContainerDied","Data":"3599f1056c197920d4bf296a8f12aa1a83b418ce08cb03b56072d18328a27d4b"} Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.579851 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3599f1056c197920d4bf296a8f12aa1a83b418ce08cb03b56072d18328a27d4b" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.579089 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c71d-account-create-update-dtbgj" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.580793 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-98sxs" event={"ID":"779f6638-c56b-462b-a574-21d3a2e08c28","Type":"ContainerDied","Data":"35628422b129d7f8b72594a56b30a303533f9f693ba8ab39c3cbf450fa6cf809"} Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.580834 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-98sxs" Dec 07 09:26:59 crc kubenswrapper[4838]: I1207 09:26:59.580840 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35628422b129d7f8b72594a56b30a303533f9f693ba8ab39c3cbf450fa6cf809" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.204032 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8285b"] Dec 07 09:27:00 crc kubenswrapper[4838]: E1207 09:27:00.204850 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e334fc-b52e-4b90-bcd6-736bdc1fcd3a" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.204870 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e334fc-b52e-4b90-bcd6-736bdc1fcd3a" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: E1207 09:27:00.204889 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe4a29d-3f3d-4cc3-884f-00f111d7d651" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.204898 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe4a29d-3f3d-4cc3-884f-00f111d7d651" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: E1207 09:27:00.204905 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd80aba1-e836-4c0f-8eba-2999e284db44" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.204914 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd80aba1-e836-4c0f-8eba-2999e284db44" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: E1207 09:27:00.204931 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779f6638-c56b-462b-a574-21d3a2e08c28" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.204938 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="779f6638-c56b-462b-a574-21d3a2e08c28" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: E1207 09:27:00.204955 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c25b096-cc08-4a86-bc3c-b5c8601b3f71" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.204962 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c25b096-cc08-4a86-bc3c-b5c8601b3f71" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: E1207 09:27:00.204978 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e90708-7b16-4279-bb16-423c616b152c" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.204986 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e90708-7b16-4279-bb16-423c616b152c" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.205211 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e334fc-b52e-4b90-bcd6-736bdc1fcd3a" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.205230 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e90708-7b16-4279-bb16-423c616b152c" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.205239 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c25b096-cc08-4a86-bc3c-b5c8601b3f71" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.205256 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe4a29d-3f3d-4cc3-884f-00f111d7d651" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.205267 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd80aba1-e836-4c0f-8eba-2999e284db44" containerName="mariadb-account-create-update" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.205280 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="779f6638-c56b-462b-a574-21d3a2e08c28" containerName="mariadb-database-create" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.205975 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.209393 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.209412 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bl7z5" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.211271 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.221651 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8285b"] Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.321275 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-config-data\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.321315 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-scripts\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.321434 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghpb\" (UniqueName: \"kubernetes.io/projected/d84318a4-61ad-4c08-928e-d06c3a99cc0b-kube-api-access-5ghpb\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.321469 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.422851 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-config-data\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.422921 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-scripts\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.423033 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghpb\" (UniqueName: \"kubernetes.io/projected/d84318a4-61ad-4c08-928e-d06c3a99cc0b-kube-api-access-5ghpb\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.423063 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.428688 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-config-data\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.429217 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-scripts\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.436587 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.441615 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghpb\" (UniqueName: \"kubernetes.io/projected/d84318a4-61ad-4c08-928e-d06c3a99cc0b-kube-api-access-5ghpb\") pod \"nova-cell0-conductor-db-sync-8285b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:00 crc kubenswrapper[4838]: I1207 09:27:00.528344 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:01 crc kubenswrapper[4838]: I1207 09:27:01.161688 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8285b"] Dec 07 09:27:01 crc kubenswrapper[4838]: W1207 09:27:01.172353 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd84318a4_61ad_4c08_928e_d06c3a99cc0b.slice/crio-32c9cbc02c44fffeffedf54cb4b1ad3d8572368840f8181c5a272ce8affe0bc8 WatchSource:0}: Error finding container 32c9cbc02c44fffeffedf54cb4b1ad3d8572368840f8181c5a272ce8affe0bc8: Status 404 returned error can't find the container with id 32c9cbc02c44fffeffedf54cb4b1ad3d8572368840f8181c5a272ce8affe0bc8 Dec 07 09:27:01 crc kubenswrapper[4838]: I1207 09:27:01.602500 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8285b" event={"ID":"d84318a4-61ad-4c08-928e-d06c3a99cc0b","Type":"ContainerStarted","Data":"32c9cbc02c44fffeffedf54cb4b1ad3d8572368840f8181c5a272ce8affe0bc8"} Dec 07 09:27:02 crc kubenswrapper[4838]: I1207 09:27:02.844985 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 07 09:27:05 crc kubenswrapper[4838]: I1207 09:27:05.647413 4838 generic.go:334] "Generic (PLEG): container finished" podID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerID="b4a6dd3a4191ba0addd0839e6003f513db1f7727a496c01ac32786e8b25662f6" exitCode=0 Dec 07 09:27:05 crc kubenswrapper[4838]: I1207 09:27:05.647445 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerDied","Data":"b4a6dd3a4191ba0addd0839e6003f513db1f7727a496c01ac32786e8b25662f6"} Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.607320 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.700557 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2270d2fe-8e5e-453a-83e1-371306ff8ef8","Type":"ContainerDied","Data":"18c166d4230430f3ba50efbced836d2111f27b08c659b2fc80900562cdb58809"} Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.700673 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.700879 4838 scope.go:117] "RemoveContainer" containerID="a2cfd138493be75e4acd0d83a1a46029012557bc8df0db42b2f8667ea813b65c" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.736588 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-sg-core-conf-yaml\") pod \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.736766 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-log-httpd\") pod \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.737384 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2270d2fe-8e5e-453a-83e1-371306ff8ef8" (UID: "2270d2fe-8e5e-453a-83e1-371306ff8ef8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.737518 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-run-httpd\") pod \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.737858 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2270d2fe-8e5e-453a-83e1-371306ff8ef8" (UID: "2270d2fe-8e5e-453a-83e1-371306ff8ef8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.737892 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-scripts\") pod \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.738031 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-combined-ca-bundle\") pod \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.738533 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-config-data\") pod \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.738610 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfc6b\" (UniqueName: \"kubernetes.io/projected/2270d2fe-8e5e-453a-83e1-371306ff8ef8-kube-api-access-tfc6b\") pod \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\" (UID: \"2270d2fe-8e5e-453a-83e1-371306ff8ef8\") " Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.739427 4838 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.739454 4838 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2270d2fe-8e5e-453a-83e1-371306ff8ef8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.741599 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-scripts" (OuterVolumeSpecName: "scripts") pod "2270d2fe-8e5e-453a-83e1-371306ff8ef8" (UID: "2270d2fe-8e5e-453a-83e1-371306ff8ef8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.741867 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2270d2fe-8e5e-453a-83e1-371306ff8ef8-kube-api-access-tfc6b" (OuterVolumeSpecName: "kube-api-access-tfc6b") pod "2270d2fe-8e5e-453a-83e1-371306ff8ef8" (UID: "2270d2fe-8e5e-453a-83e1-371306ff8ef8"). InnerVolumeSpecName "kube-api-access-tfc6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.759517 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2270d2fe-8e5e-453a-83e1-371306ff8ef8" (UID: "2270d2fe-8e5e-453a-83e1-371306ff8ef8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.772212 4838 scope.go:117] "RemoveContainer" containerID="a876029f334ad4701134b0979763a72f1b1ef0693032afa9fd785f733086db91" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.795202 4838 scope.go:117] "RemoveContainer" containerID="e6ed6db9299997cec854112736dad69b194ae457392c3fedb07261b3498ce8e4" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.820076 4838 scope.go:117] "RemoveContainer" containerID="b4a6dd3a4191ba0addd0839e6003f513db1f7727a496c01ac32786e8b25662f6" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.822362 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2270d2fe-8e5e-453a-83e1-371306ff8ef8" (UID: "2270d2fe-8e5e-453a-83e1-371306ff8ef8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.837651 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-config-data" (OuterVolumeSpecName: "config-data") pod "2270d2fe-8e5e-453a-83e1-371306ff8ef8" (UID: "2270d2fe-8e5e-453a-83e1-371306ff8ef8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.840679 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.840703 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.840712 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.840722 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfc6b\" (UniqueName: \"kubernetes.io/projected/2270d2fe-8e5e-453a-83e1-371306ff8ef8-kube-api-access-tfc6b\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:10 crc kubenswrapper[4838]: I1207 09:27:10.840733 4838 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2270d2fe-8e5e-453a-83e1-371306ff8ef8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.057340 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.065652 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091184 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:27:11 crc kubenswrapper[4838]: E1207 09:27:11.091519 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-notification-agent" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091537 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-notification-agent" Dec 07 09:27:11 crc kubenswrapper[4838]: E1207 09:27:11.091556 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="sg-core" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091562 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="sg-core" Dec 07 09:27:11 crc kubenswrapper[4838]: E1207 09:27:11.091577 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="proxy-httpd" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091583 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="proxy-httpd" Dec 07 09:27:11 crc kubenswrapper[4838]: E1207 09:27:11.091598 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-central-agent" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091604 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-central-agent" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091761 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-central-agent" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091773 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="proxy-httpd" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091781 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="ceilometer-notification-agent" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.091790 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" containerName="sg-core" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.100892 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.106675 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.107011 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.107157 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.126864 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153204 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-config-data\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153322 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-log-httpd\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153416 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153543 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-run-httpd\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153584 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153612 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153643 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwfsz\" (UniqueName: \"kubernetes.io/projected/693d763b-81c2-4279-9160-668a1819e01b-kube-api-access-xwfsz\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.153757 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-scripts\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.254949 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-run-httpd\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.254994 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.255012 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.255031 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwfsz\" (UniqueName: \"kubernetes.io/projected/693d763b-81c2-4279-9160-668a1819e01b-kube-api-access-xwfsz\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.255061 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-scripts\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.255111 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-config-data\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.255144 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-log-httpd\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.255166 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.257290 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-run-httpd\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.257591 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-log-httpd\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.258427 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-scripts\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.258992 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-config-data\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.259658 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.259674 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.260313 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.275431 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwfsz\" (UniqueName: \"kubernetes.io/projected/693d763b-81c2-4279-9160-668a1819e01b-kube-api-access-xwfsz\") pod \"ceilometer-0\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.431671 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.627421 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2270d2fe-8e5e-453a-83e1-371306ff8ef8" path="/var/lib/kubelet/pods/2270d2fe-8e5e-453a-83e1-371306ff8ef8/volumes" Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.719851 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8285b" event={"ID":"d84318a4-61ad-4c08-928e-d06c3a99cc0b","Type":"ContainerStarted","Data":"4eb99b3245d0e86a5b8360cd9b492896486e98c4d81604616e45226be633955c"} Dec 07 09:27:11 crc kubenswrapper[4838]: I1207 09:27:11.742867 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-8285b" podStartSLOduration=2.276045439 podStartE2EDuration="11.742848945s" podCreationTimestamp="2025-12-07 09:27:00 +0000 UTC" firstStartedPulling="2025-12-07 09:27:01.174316705 +0000 UTC m=+1237.881635722" lastFinishedPulling="2025-12-07 09:27:10.641120211 +0000 UTC m=+1247.348439228" observedRunningTime="2025-12-07 09:27:11.735197403 +0000 UTC m=+1248.442516430" watchObservedRunningTime="2025-12-07 09:27:11.742848945 +0000 UTC m=+1248.450167982" Dec 07 09:27:12 crc kubenswrapper[4838]: I1207 09:27:12.013508 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:27:12 crc kubenswrapper[4838]: W1207 09:27:12.020558 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod693d763b_81c2_4279_9160_668a1819e01b.slice/crio-739113d5b7a9057093649e9e0cec4617886259f096d663e62cb127c39e29bdd7 WatchSource:0}: Error finding container 739113d5b7a9057093649e9e0cec4617886259f096d663e62cb127c39e29bdd7: Status 404 returned error can't find the container with id 739113d5b7a9057093649e9e0cec4617886259f096d663e62cb127c39e29bdd7 Dec 07 09:27:12 crc kubenswrapper[4838]: I1207 09:27:12.729983 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerStarted","Data":"087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6"} Dec 07 09:27:12 crc kubenswrapper[4838]: I1207 09:27:12.730322 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerStarted","Data":"739113d5b7a9057093649e9e0cec4617886259f096d663e62cb127c39e29bdd7"} Dec 07 09:27:13 crc kubenswrapper[4838]: I1207 09:27:13.745402 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerStarted","Data":"ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924"} Dec 07 09:27:14 crc kubenswrapper[4838]: I1207 09:27:14.755692 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerStarted","Data":"0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d"} Dec 07 09:27:15 crc kubenswrapper[4838]: I1207 09:27:15.765037 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerStarted","Data":"eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806"} Dec 07 09:27:15 crc kubenswrapper[4838]: I1207 09:27:15.765966 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 09:27:15 crc kubenswrapper[4838]: I1207 09:27:15.787359 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.874206388 podStartE2EDuration="4.787337955s" podCreationTimestamp="2025-12-07 09:27:11 +0000 UTC" firstStartedPulling="2025-12-07 09:27:12.025791473 +0000 UTC m=+1248.733110490" lastFinishedPulling="2025-12-07 09:27:14.93892304 +0000 UTC m=+1251.646242057" observedRunningTime="2025-12-07 09:27:15.782768185 +0000 UTC m=+1252.490087202" watchObservedRunningTime="2025-12-07 09:27:15.787337955 +0000 UTC m=+1252.494656972" Dec 07 09:27:21 crc kubenswrapper[4838]: I1207 09:27:21.818634 4838 generic.go:334] "Generic (PLEG): container finished" podID="d84318a4-61ad-4c08-928e-d06c3a99cc0b" containerID="4eb99b3245d0e86a5b8360cd9b492896486e98c4d81604616e45226be633955c" exitCode=0 Dec 07 09:27:21 crc kubenswrapper[4838]: I1207 09:27:21.818739 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8285b" event={"ID":"d84318a4-61ad-4c08-928e-d06c3a99cc0b","Type":"ContainerDied","Data":"4eb99b3245d0e86a5b8360cd9b492896486e98c4d81604616e45226be633955c"} Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.110669 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.287329 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-config-data\") pod \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.287420 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ghpb\" (UniqueName: \"kubernetes.io/projected/d84318a4-61ad-4c08-928e-d06c3a99cc0b-kube-api-access-5ghpb\") pod \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.287461 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-scripts\") pod \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.287513 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-combined-ca-bundle\") pod \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\" (UID: \"d84318a4-61ad-4c08-928e-d06c3a99cc0b\") " Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.294033 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-scripts" (OuterVolumeSpecName: "scripts") pod "d84318a4-61ad-4c08-928e-d06c3a99cc0b" (UID: "d84318a4-61ad-4c08-928e-d06c3a99cc0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.294750 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84318a4-61ad-4c08-928e-d06c3a99cc0b-kube-api-access-5ghpb" (OuterVolumeSpecName: "kube-api-access-5ghpb") pod "d84318a4-61ad-4c08-928e-d06c3a99cc0b" (UID: "d84318a4-61ad-4c08-928e-d06c3a99cc0b"). InnerVolumeSpecName "kube-api-access-5ghpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.331040 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d84318a4-61ad-4c08-928e-d06c3a99cc0b" (UID: "d84318a4-61ad-4c08-928e-d06c3a99cc0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.336483 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-config-data" (OuterVolumeSpecName: "config-data") pod "d84318a4-61ad-4c08-928e-d06c3a99cc0b" (UID: "d84318a4-61ad-4c08-928e-d06c3a99cc0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.389339 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ghpb\" (UniqueName: \"kubernetes.io/projected/d84318a4-61ad-4c08-928e-d06c3a99cc0b-kube-api-access-5ghpb\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.389382 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.389397 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.389410 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84318a4-61ad-4c08-928e-d06c3a99cc0b-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.840117 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8285b" event={"ID":"d84318a4-61ad-4c08-928e-d06c3a99cc0b","Type":"ContainerDied","Data":"32c9cbc02c44fffeffedf54cb4b1ad3d8572368840f8181c5a272ce8affe0bc8"} Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.840563 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32c9cbc02c44fffeffedf54cb4b1ad3d8572368840f8181c5a272ce8affe0bc8" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.840241 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8285b" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.949136 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 07 09:27:23 crc kubenswrapper[4838]: E1207 09:27:23.949958 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84318a4-61ad-4c08-928e-d06c3a99cc0b" containerName="nova-cell0-conductor-db-sync" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.950047 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84318a4-61ad-4c08-928e-d06c3a99cc0b" containerName="nova-cell0-conductor-db-sync" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.950362 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84318a4-61ad-4c08-928e-d06c3a99cc0b" containerName="nova-cell0-conductor-db-sync" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.951103 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.954457 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bl7z5" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.954568 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 07 09:27:23 crc kubenswrapper[4838]: I1207 09:27:23.971115 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.100510 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l2fq\" (UniqueName: \"kubernetes.io/projected/485d8af5-c6b8-4d04-be66-a3a421f86d49-kube-api-access-5l2fq\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.100597 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/485d8af5-c6b8-4d04-be66-a3a421f86d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.100749 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/485d8af5-c6b8-4d04-be66-a3a421f86d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.201904 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/485d8af5-c6b8-4d04-be66-a3a421f86d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.201970 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/485d8af5-c6b8-4d04-be66-a3a421f86d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.202085 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l2fq\" (UniqueName: \"kubernetes.io/projected/485d8af5-c6b8-4d04-be66-a3a421f86d49-kube-api-access-5l2fq\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.209751 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/485d8af5-c6b8-4d04-be66-a3a421f86d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.214768 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/485d8af5-c6b8-4d04-be66-a3a421f86d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.221445 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l2fq\" (UniqueName: \"kubernetes.io/projected/485d8af5-c6b8-4d04-be66-a3a421f86d49-kube-api-access-5l2fq\") pod \"nova-cell0-conductor-0\" (UID: \"485d8af5-c6b8-4d04-be66-a3a421f86d49\") " pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.268508 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.720715 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 07 09:27:24 crc kubenswrapper[4838]: I1207 09:27:24.851159 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"485d8af5-c6b8-4d04-be66-a3a421f86d49","Type":"ContainerStarted","Data":"1c2ef543ba6dec9f1cd4abafecce89e5208153a55518fd4094dfb51d1ef40795"} Dec 07 09:27:25 crc kubenswrapper[4838]: I1207 09:27:25.865447 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"485d8af5-c6b8-4d04-be66-a3a421f86d49","Type":"ContainerStarted","Data":"8a26b4d3a3b80cf173d1401608ae5e9b30e7205431859fafcfa11f97fe503a42"} Dec 07 09:27:25 crc kubenswrapper[4838]: I1207 09:27:25.867353 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.305192 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.324754 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=6.324737874 podStartE2EDuration="6.324737874s" podCreationTimestamp="2025-12-07 09:27:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:25.891554716 +0000 UTC m=+1262.598873753" watchObservedRunningTime="2025-12-07 09:27:29.324737874 +0000 UTC m=+1266.032056891" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.755617 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-v59wv"] Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.757190 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.759656 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.760460 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.781863 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-v59wv"] Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.811085 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-config-data\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.811491 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-scripts\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.811752 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.811962 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nx5h\" (UniqueName: \"kubernetes.io/projected/63fb1e51-5812-4adf-9e44-a7f4cd3349da-kube-api-access-8nx5h\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.932320 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-config-data\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.932397 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-scripts\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.932440 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.932473 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nx5h\" (UniqueName: \"kubernetes.io/projected/63fb1e51-5812-4adf-9e44-a7f4cd3349da-kube-api-access-8nx5h\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.944439 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-scripts\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.945176 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.954676 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-config-data\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:29 crc kubenswrapper[4838]: I1207 09:27:29.964210 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nx5h\" (UniqueName: \"kubernetes.io/projected/63fb1e51-5812-4adf-9e44-a7f4cd3349da-kube-api-access-8nx5h\") pod \"nova-cell0-cell-mapping-v59wv\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.028746 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.030728 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.045890 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.047001 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.047338 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.058066 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.071130 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.093185 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.131407 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.172507 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncwh8\" (UniqueName: \"kubernetes.io/projected/81adc2d0-f8c4-4b96-b459-9a446d468111-kube-api-access-ncwh8\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.172596 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.172636 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-config-data\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.172662 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.172702 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxt96\" (UniqueName: \"kubernetes.io/projected/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-kube-api-access-cxt96\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.172742 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-logs\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.172803 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.198808 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.200322 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.209929 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.212286 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.273890 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxt96\" (UniqueName: \"kubernetes.io/projected/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-kube-api-access-cxt96\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.273962 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-logs\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.273986 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ab174e0-f973-4f02-8d13-72ea13c9dd25-logs\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274035 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274060 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274101 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncwh8\" (UniqueName: \"kubernetes.io/projected/81adc2d0-f8c4-4b96-b459-9a446d468111-kube-api-access-ncwh8\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274128 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274144 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl4gh\" (UniqueName: \"kubernetes.io/projected/6ab174e0-f973-4f02-8d13-72ea13c9dd25-kube-api-access-jl4gh\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274170 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-config-data\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274189 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274208 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-config-data\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.274907 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-logs\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.284455 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.285417 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.287071 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-config-data\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.308010 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.320089 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncwh8\" (UniqueName: \"kubernetes.io/projected/81adc2d0-f8c4-4b96-b459-9a446d468111-kube-api-access-ncwh8\") pod \"nova-cell1-novncproxy-0\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.331169 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxt96\" (UniqueName: \"kubernetes.io/projected/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-kube-api-access-cxt96\") pod \"nova-api-0\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.369544 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.395474 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.398965 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ab174e0-f973-4f02-8d13-72ea13c9dd25-logs\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.399057 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.399129 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl4gh\" (UniqueName: \"kubernetes.io/projected/6ab174e0-f973-4f02-8d13-72ea13c9dd25-kube-api-access-jl4gh\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.399167 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-config-data\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.413722 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-7f9tn"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.415242 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.428678 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ab174e0-f973-4f02-8d13-72ea13c9dd25-logs\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.438212 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-config-data\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.445649 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.455535 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.456179 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl4gh\" (UniqueName: \"kubernetes.io/projected/6ab174e0-f973-4f02-8d13-72ea13c9dd25-kube-api-access-jl4gh\") pod \"nova-metadata-0\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.456645 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.461887 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.497513 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-7f9tn"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.503315 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-sb\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.503398 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-config\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.503489 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-dns-svc\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.503522 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-nb\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.503564 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct6mw\" (UniqueName: \"kubernetes.io/projected/81515be6-4b04-4629-9ff8-37362ab0a313-kube-api-access-ct6mw\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.545913 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.552269 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.605924 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-config-data\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.606221 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct6mw\" (UniqueName: \"kubernetes.io/projected/81515be6-4b04-4629-9ff8-37362ab0a313-kube-api-access-ct6mw\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.606258 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-sb\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.606285 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.606334 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-config\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.606352 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltrst\" (UniqueName: \"kubernetes.io/projected/bc296916-0117-4546-ae05-f9860fdf47b4-kube-api-access-ltrst\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.606418 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-dns-svc\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.606455 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-nb\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.607259 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-nb\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.608171 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-sb\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.608622 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-config\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.608737 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-dns-svc\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.638445 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct6mw\" (UniqueName: \"kubernetes.io/projected/81515be6-4b04-4629-9ff8-37362ab0a313-kube-api-access-ct6mw\") pod \"dnsmasq-dns-55cbc7dc9c-7f9tn\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.710413 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-config-data\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.710493 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.710570 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltrst\" (UniqueName: \"kubernetes.io/projected/bc296916-0117-4546-ae05-f9860fdf47b4-kube-api-access-ltrst\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.715337 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.722142 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-config-data\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.734059 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltrst\" (UniqueName: \"kubernetes.io/projected/bc296916-0117-4546-ae05-f9860fdf47b4-kube-api-access-ltrst\") pod \"nova-scheduler-0\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.803123 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.826235 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.844744 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-v59wv"] Dec 07 09:27:30 crc kubenswrapper[4838]: W1207 09:27:30.856171 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63fb1e51_5812_4adf_9e44_a7f4cd3349da.slice/crio-e4485f84475dcdf96479dc1014e837b45f5fe783bd52c35aca3786e272ffc669 WatchSource:0}: Error finding container e4485f84475dcdf96479dc1014e837b45f5fe783bd52c35aca3786e272ffc669: Status 404 returned error can't find the container with id e4485f84475dcdf96479dc1014e837b45f5fe783bd52c35aca3786e272ffc669 Dec 07 09:27:30 crc kubenswrapper[4838]: I1207 09:27:30.960806 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v59wv" event={"ID":"63fb1e51-5812-4adf-9e44-a7f4cd3349da","Type":"ContainerStarted","Data":"e4485f84475dcdf96479dc1014e837b45f5fe783bd52c35aca3786e272ffc669"} Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.062671 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.117533 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.228772 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h7vd6"] Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.232877 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.245070 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.245110 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.247582 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h7vd6"] Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.255858 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-scripts\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.255932 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.255996 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-config-data\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.256076 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzwvj\" (UniqueName: \"kubernetes.io/projected/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-kube-api-access-bzwvj\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.308588 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.358300 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-scripts\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.358366 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.358434 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-config-data\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.358488 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzwvj\" (UniqueName: \"kubernetes.io/projected/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-kube-api-access-bzwvj\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.364477 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.364898 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-scripts\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.369076 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-config-data\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.380940 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzwvj\" (UniqueName: \"kubernetes.io/projected/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-kube-api-access-bzwvj\") pod \"nova-cell1-conductor-db-sync-h7vd6\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.437909 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.573860 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.598426 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-7f9tn"] Dec 07 09:27:31 crc kubenswrapper[4838]: W1207 09:27:31.599408 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81515be6_4b04_4629_9ff8_37362ab0a313.slice/crio-4cbc7c21fbef34268e6cb1151b33366027d44d2ef25641be81fd256f43acc38b WatchSource:0}: Error finding container 4cbc7c21fbef34268e6cb1151b33366027d44d2ef25641be81fd256f43acc38b: Status 404 returned error can't find the container with id 4cbc7c21fbef34268e6cb1151b33366027d44d2ef25641be81fd256f43acc38b Dec 07 09:27:31 crc kubenswrapper[4838]: I1207 09:27:31.983057 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v59wv" event={"ID":"63fb1e51-5812-4adf-9e44-a7f4cd3349da","Type":"ContainerStarted","Data":"7ddff9e775deed21a1736b371cbedd551de97ea5d9f8105586db67c68865d63c"} Dec 07 09:27:32 crc kubenswrapper[4838]: I1207 09:27:32.005560 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ab174e0-f973-4f02-8d13-72ea13c9dd25","Type":"ContainerStarted","Data":"67963efbbcbe2488a8e6b3b92c1b891d7a7ee299b0d2f9d481085d9946db428c"} Dec 07 09:27:32 crc kubenswrapper[4838]: I1207 09:27:32.017730 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-v59wv" podStartSLOduration=3.017706063 podStartE2EDuration="3.017706063s" podCreationTimestamp="2025-12-07 09:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:32.010378 +0000 UTC m=+1268.717697017" watchObservedRunningTime="2025-12-07 09:27:32.017706063 +0000 UTC m=+1268.725025080" Dec 07 09:27:32 crc kubenswrapper[4838]: I1207 09:27:32.020670 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8178ebbf-dd4e-4456-acfd-63ebbd26a95b","Type":"ContainerStarted","Data":"627b6fa374d4e24e25e61c607bfa4c7cfb954a3f84765d9e2992bb3e12f1f771"} Dec 07 09:27:32 crc kubenswrapper[4838]: I1207 09:27:32.021885 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"81adc2d0-f8c4-4b96-b459-9a446d468111","Type":"ContainerStarted","Data":"32b02e1c454cc3a9dcdba84cae6ac2d9e97eebd72dec405fcf1d3dafe54ea213"} Dec 07 09:27:32 crc kubenswrapper[4838]: I1207 09:27:32.040994 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" event={"ID":"81515be6-4b04-4629-9ff8-37362ab0a313","Type":"ContainerStarted","Data":"4cbc7c21fbef34268e6cb1151b33366027d44d2ef25641be81fd256f43acc38b"} Dec 07 09:27:32 crc kubenswrapper[4838]: I1207 09:27:32.048588 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bc296916-0117-4546-ae05-f9860fdf47b4","Type":"ContainerStarted","Data":"a11b73967ca60c5e700fdcf981ae4153cc5e621810ae59b136b0085bc7f386cc"} Dec 07 09:27:32 crc kubenswrapper[4838]: I1207 09:27:32.183866 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h7vd6"] Dec 07 09:27:33 crc kubenswrapper[4838]: I1207 09:27:33.059508 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" event={"ID":"99b4af2f-6fae-4a99-a9ee-8447c2a00df3","Type":"ContainerStarted","Data":"f1331b139ed990a28aebd176d4684b6187ee3235742df54bf6b28726119d4445"} Dec 07 09:27:33 crc kubenswrapper[4838]: I1207 09:27:33.060040 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" event={"ID":"99b4af2f-6fae-4a99-a9ee-8447c2a00df3","Type":"ContainerStarted","Data":"6f3db32ad1f2f9d39c5febcda2399b34c1793c43c08200479a46888f0af95092"} Dec 07 09:27:33 crc kubenswrapper[4838]: I1207 09:27:33.082984 4838 generic.go:334] "Generic (PLEG): container finished" podID="81515be6-4b04-4629-9ff8-37362ab0a313" containerID="edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7" exitCode=0 Dec 07 09:27:33 crc kubenswrapper[4838]: I1207 09:27:33.083068 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" event={"ID":"81515be6-4b04-4629-9ff8-37362ab0a313","Type":"ContainerDied","Data":"edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7"} Dec 07 09:27:33 crc kubenswrapper[4838]: I1207 09:27:33.103859 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" podStartSLOduration=2.103841907 podStartE2EDuration="2.103841907s" podCreationTimestamp="2025-12-07 09:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:33.07696729 +0000 UTC m=+1269.784286307" watchObservedRunningTime="2025-12-07 09:27:33.103841907 +0000 UTC m=+1269.811160924" Dec 07 09:27:34 crc kubenswrapper[4838]: I1207 09:27:34.899866 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:27:34 crc kubenswrapper[4838]: I1207 09:27:34.908237 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:36 crc kubenswrapper[4838]: I1207 09:27:36.130681 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ab174e0-f973-4f02-8d13-72ea13c9dd25","Type":"ContainerStarted","Data":"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa"} Dec 07 09:27:36 crc kubenswrapper[4838]: I1207 09:27:36.134567 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8178ebbf-dd4e-4456-acfd-63ebbd26a95b","Type":"ContainerStarted","Data":"48f87d65dc3b784f42b5961225c18c0feaa77fb4e6fc5c2c73b3651071e55b9a"} Dec 07 09:27:36 crc kubenswrapper[4838]: I1207 09:27:36.136444 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"81adc2d0-f8c4-4b96-b459-9a446d468111","Type":"ContainerStarted","Data":"29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf"} Dec 07 09:27:36 crc kubenswrapper[4838]: I1207 09:27:36.136567 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="81adc2d0-f8c4-4b96-b459-9a446d468111" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf" gracePeriod=30 Dec 07 09:27:36 crc kubenswrapper[4838]: I1207 09:27:36.138803 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bc296916-0117-4546-ae05-f9860fdf47b4","Type":"ContainerStarted","Data":"9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19"} Dec 07 09:27:36 crc kubenswrapper[4838]: I1207 09:27:36.154683 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.7376887500000002 podStartE2EDuration="7.154666093s" podCreationTimestamp="2025-12-07 09:27:29 +0000 UTC" firstStartedPulling="2025-12-07 09:27:31.163255299 +0000 UTC m=+1267.870574306" lastFinishedPulling="2025-12-07 09:27:35.580232632 +0000 UTC m=+1272.287551649" observedRunningTime="2025-12-07 09:27:36.154116169 +0000 UTC m=+1272.861435186" watchObservedRunningTime="2025-12-07 09:27:36.154666093 +0000 UTC m=+1272.861985110" Dec 07 09:27:36 crc kubenswrapper[4838]: I1207 09:27:36.182260 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.039254018 podStartE2EDuration="6.182239778s" podCreationTimestamp="2025-12-07 09:27:30 +0000 UTC" firstStartedPulling="2025-12-07 09:27:31.438956767 +0000 UTC m=+1268.146275784" lastFinishedPulling="2025-12-07 09:27:35.581942517 +0000 UTC m=+1272.289261544" observedRunningTime="2025-12-07 09:27:36.180179654 +0000 UTC m=+1272.887498671" watchObservedRunningTime="2025-12-07 09:27:36.182239778 +0000 UTC m=+1272.889558795" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.159688 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ab174e0-f973-4f02-8d13-72ea13c9dd25","Type":"ContainerStarted","Data":"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe"} Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.159886 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-log" containerID="cri-o://a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa" gracePeriod=30 Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.160499 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-metadata" containerID="cri-o://bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe" gracePeriod=30 Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.163699 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8178ebbf-dd4e-4456-acfd-63ebbd26a95b","Type":"ContainerStarted","Data":"996ff1647f93dc62377fe7f0c3b4d9f442b1a28f5e1d6642922ccf3f85c80916"} Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.178389 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" event={"ID":"81515be6-4b04-4629-9ff8-37362ab0a313","Type":"ContainerStarted","Data":"3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2"} Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.178451 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.187209 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.916952373 podStartE2EDuration="7.187192259s" podCreationTimestamp="2025-12-07 09:27:30 +0000 UTC" firstStartedPulling="2025-12-07 09:27:31.314914526 +0000 UTC m=+1268.022233543" lastFinishedPulling="2025-12-07 09:27:35.585154412 +0000 UTC m=+1272.292473429" observedRunningTime="2025-12-07 09:27:37.185107494 +0000 UTC m=+1273.892426511" watchObservedRunningTime="2025-12-07 09:27:37.187192259 +0000 UTC m=+1273.894511276" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.227977 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" podStartSLOduration=7.22795484 podStartE2EDuration="7.22795484s" podCreationTimestamp="2025-12-07 09:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:37.225892986 +0000 UTC m=+1273.933212003" watchObservedRunningTime="2025-12-07 09:27:37.22795484 +0000 UTC m=+1273.935273857" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.272526 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.778443693 podStartE2EDuration="8.272503922s" podCreationTimestamp="2025-12-07 09:27:29 +0000 UTC" firstStartedPulling="2025-12-07 09:27:31.085951327 +0000 UTC m=+1267.793270344" lastFinishedPulling="2025-12-07 09:27:35.580011556 +0000 UTC m=+1272.287330573" observedRunningTime="2025-12-07 09:27:37.25002054 +0000 UTC m=+1273.957339577" watchObservedRunningTime="2025-12-07 09:27:37.272503922 +0000 UTC m=+1273.979822939" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.811480 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.917417 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl4gh\" (UniqueName: \"kubernetes.io/projected/6ab174e0-f973-4f02-8d13-72ea13c9dd25-kube-api-access-jl4gh\") pod \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.917688 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ab174e0-f973-4f02-8d13-72ea13c9dd25-logs\") pod \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.917776 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-config-data\") pod \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.917792 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-combined-ca-bundle\") pod \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\" (UID: \"6ab174e0-f973-4f02-8d13-72ea13c9dd25\") " Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.918346 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ab174e0-f973-4f02-8d13-72ea13c9dd25-logs" (OuterVolumeSpecName: "logs") pod "6ab174e0-f973-4f02-8d13-72ea13c9dd25" (UID: "6ab174e0-f973-4f02-8d13-72ea13c9dd25"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.924143 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ab174e0-f973-4f02-8d13-72ea13c9dd25-kube-api-access-jl4gh" (OuterVolumeSpecName: "kube-api-access-jl4gh") pod "6ab174e0-f973-4f02-8d13-72ea13c9dd25" (UID: "6ab174e0-f973-4f02-8d13-72ea13c9dd25"). InnerVolumeSpecName "kube-api-access-jl4gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.943020 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ab174e0-f973-4f02-8d13-72ea13c9dd25" (UID: "6ab174e0-f973-4f02-8d13-72ea13c9dd25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:37 crc kubenswrapper[4838]: I1207 09:27:37.949922 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-config-data" (OuterVolumeSpecName: "config-data") pod "6ab174e0-f973-4f02-8d13-72ea13c9dd25" (UID: "6ab174e0-f973-4f02-8d13-72ea13c9dd25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.020096 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl4gh\" (UniqueName: \"kubernetes.io/projected/6ab174e0-f973-4f02-8d13-72ea13c9dd25-kube-api-access-jl4gh\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.020138 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ab174e0-f973-4f02-8d13-72ea13c9dd25-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.020153 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.020168 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ab174e0-f973-4f02-8d13-72ea13c9dd25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.187413 4838 generic.go:334] "Generic (PLEG): container finished" podID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerID="bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe" exitCode=0 Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.187443 4838 generic.go:334] "Generic (PLEG): container finished" podID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerID="a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa" exitCode=143 Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.188298 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ab174e0-f973-4f02-8d13-72ea13c9dd25","Type":"ContainerDied","Data":"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe"} Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.188325 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ab174e0-f973-4f02-8d13-72ea13c9dd25","Type":"ContainerDied","Data":"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa"} Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.188335 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ab174e0-f973-4f02-8d13-72ea13c9dd25","Type":"ContainerDied","Data":"67963efbbcbe2488a8e6b3b92c1b891d7a7ee299b0d2f9d481085d9946db428c"} Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.188336 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.188351 4838 scope.go:117] "RemoveContainer" containerID="bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.211725 4838 scope.go:117] "RemoveContainer" containerID="a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.229520 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.253294 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.256019 4838 scope.go:117] "RemoveContainer" containerID="bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe" Dec 07 09:27:38 crc kubenswrapper[4838]: E1207 09:27:38.257289 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe\": container with ID starting with bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe not found: ID does not exist" containerID="bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.257331 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe"} err="failed to get container status \"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe\": rpc error: code = NotFound desc = could not find container \"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe\": container with ID starting with bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe not found: ID does not exist" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.257357 4838 scope.go:117] "RemoveContainer" containerID="a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa" Dec 07 09:27:38 crc kubenswrapper[4838]: E1207 09:27:38.259592 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa\": container with ID starting with a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa not found: ID does not exist" containerID="a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.259629 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa"} err="failed to get container status \"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa\": rpc error: code = NotFound desc = could not find container \"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa\": container with ID starting with a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa not found: ID does not exist" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.259650 4838 scope.go:117] "RemoveContainer" containerID="bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.260926 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe"} err="failed to get container status \"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe\": rpc error: code = NotFound desc = could not find container \"bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe\": container with ID starting with bc15cdcf9ce53229116ca504b12229160a2cb9d5a746e25fd7e98bb4d1dd6cbe not found: ID does not exist" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.260980 4838 scope.go:117] "RemoveContainer" containerID="a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.264984 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa"} err="failed to get container status \"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa\": rpc error: code = NotFound desc = could not find container \"a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa\": container with ID starting with a72be234409f9bcc86a074c2bd30d399111395116611798046f226b1aabbcefa not found: ID does not exist" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.268507 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:38 crc kubenswrapper[4838]: E1207 09:27:38.269053 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-metadata" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.269075 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-metadata" Dec 07 09:27:38 crc kubenswrapper[4838]: E1207 09:27:38.269101 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-log" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.269109 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-log" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.269289 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-log" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.269308 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" containerName="nova-metadata-metadata" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.270530 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.281602 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.282119 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.282306 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.324626 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.324719 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-config-data\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.324745 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d55jf\" (UniqueName: \"kubernetes.io/projected/72c05589-25ab-449e-b591-8865398f5431-kube-api-access-d55jf\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.324786 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c05589-25ab-449e-b591-8865398f5431-logs\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.325188 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.427099 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.427188 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-config-data\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.427211 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d55jf\" (UniqueName: \"kubernetes.io/projected/72c05589-25ab-449e-b591-8865398f5431-kube-api-access-d55jf\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.427245 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c05589-25ab-449e-b591-8865398f5431-logs\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.427273 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.428424 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c05589-25ab-449e-b591-8865398f5431-logs\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.431549 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.437954 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-config-data\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.438534 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.452409 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d55jf\" (UniqueName: \"kubernetes.io/projected/72c05589-25ab-449e-b591-8865398f5431-kube-api-access-d55jf\") pod \"nova-metadata-0\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " pod="openstack/nova-metadata-0" Dec 07 09:27:38 crc kubenswrapper[4838]: I1207 09:27:38.601941 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:39 crc kubenswrapper[4838]: I1207 09:27:39.159352 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:39 crc kubenswrapper[4838]: I1207 09:27:39.211659 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"72c05589-25ab-449e-b591-8865398f5431","Type":"ContainerStarted","Data":"e4839ddb9a9aacdea576243857b2a2ac10b4fd04db7925783195458fe18dce96"} Dec 07 09:27:39 crc kubenswrapper[4838]: I1207 09:27:39.624335 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ab174e0-f973-4f02-8d13-72ea13c9dd25" path="/var/lib/kubelet/pods/6ab174e0-f973-4f02-8d13-72ea13c9dd25/volumes" Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.225591 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"72c05589-25ab-449e-b591-8865398f5431","Type":"ContainerStarted","Data":"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02"} Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.225636 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"72c05589-25ab-449e-b591-8865398f5431","Type":"ContainerStarted","Data":"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436"} Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.259597 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.259574982 podStartE2EDuration="2.259574982s" podCreationTimestamp="2025-12-07 09:27:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:40.258095833 +0000 UTC m=+1276.965414860" watchObservedRunningTime="2025-12-07 09:27:40.259574982 +0000 UTC m=+1276.966893999" Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.371047 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.371126 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.396350 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.826897 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.827278 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 07 09:27:40 crc kubenswrapper[4838]: I1207 09:27:40.882482 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.235737 4838 generic.go:334] "Generic (PLEG): container finished" podID="63fb1e51-5812-4adf-9e44-a7f4cd3349da" containerID="7ddff9e775deed21a1736b371cbedd551de97ea5d9f8105586db67c68865d63c" exitCode=0 Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.235858 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v59wv" event={"ID":"63fb1e51-5812-4adf-9e44-a7f4cd3349da","Type":"ContainerDied","Data":"7ddff9e775deed21a1736b371cbedd551de97ea5d9f8105586db67c68865d63c"} Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.238267 4838 generic.go:334] "Generic (PLEG): container finished" podID="99b4af2f-6fae-4a99-a9ee-8447c2a00df3" containerID="f1331b139ed990a28aebd176d4684b6187ee3235742df54bf6b28726119d4445" exitCode=0 Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.239108 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" event={"ID":"99b4af2f-6fae-4a99-a9ee-8447c2a00df3","Type":"ContainerDied","Data":"f1331b139ed990a28aebd176d4684b6187ee3235742df54bf6b28726119d4445"} Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.270434 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.453257 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.165:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.455322 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.165:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 09:27:41 crc kubenswrapper[4838]: I1207 09:27:41.465037 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.717853 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.731472 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837094 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzwvj\" (UniqueName: \"kubernetes.io/projected/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-kube-api-access-bzwvj\") pod \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837135 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-scripts\") pod \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837197 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-scripts\") pod \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837237 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nx5h\" (UniqueName: \"kubernetes.io/projected/63fb1e51-5812-4adf-9e44-a7f4cd3349da-kube-api-access-8nx5h\") pod \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837274 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-combined-ca-bundle\") pod \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837357 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-config-data\") pod \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837400 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-config-data\") pod \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\" (UID: \"63fb1e51-5812-4adf-9e44-a7f4cd3349da\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.837439 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-combined-ca-bundle\") pod \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\" (UID: \"99b4af2f-6fae-4a99-a9ee-8447c2a00df3\") " Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.844730 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-scripts" (OuterVolumeSpecName: "scripts") pod "63fb1e51-5812-4adf-9e44-a7f4cd3349da" (UID: "63fb1e51-5812-4adf-9e44-a7f4cd3349da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.845048 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-kube-api-access-bzwvj" (OuterVolumeSpecName: "kube-api-access-bzwvj") pod "99b4af2f-6fae-4a99-a9ee-8447c2a00df3" (UID: "99b4af2f-6fae-4a99-a9ee-8447c2a00df3"). InnerVolumeSpecName "kube-api-access-bzwvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.853452 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63fb1e51-5812-4adf-9e44-a7f4cd3349da-kube-api-access-8nx5h" (OuterVolumeSpecName: "kube-api-access-8nx5h") pod "63fb1e51-5812-4adf-9e44-a7f4cd3349da" (UID: "63fb1e51-5812-4adf-9e44-a7f4cd3349da"). InnerVolumeSpecName "kube-api-access-8nx5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.855634 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-scripts" (OuterVolumeSpecName: "scripts") pod "99b4af2f-6fae-4a99-a9ee-8447c2a00df3" (UID: "99b4af2f-6fae-4a99-a9ee-8447c2a00df3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.888276 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63fb1e51-5812-4adf-9e44-a7f4cd3349da" (UID: "63fb1e51-5812-4adf-9e44-a7f4cd3349da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.904259 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99b4af2f-6fae-4a99-a9ee-8447c2a00df3" (UID: "99b4af2f-6fae-4a99-a9ee-8447c2a00df3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.910445 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-config-data" (OuterVolumeSpecName: "config-data") pod "99b4af2f-6fae-4a99-a9ee-8447c2a00df3" (UID: "99b4af2f-6fae-4a99-a9ee-8447c2a00df3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.910584 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-config-data" (OuterVolumeSpecName: "config-data") pod "63fb1e51-5812-4adf-9e44-a7f4cd3349da" (UID: "63fb1e51-5812-4adf-9e44-a7f4cd3349da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939578 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939675 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939689 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzwvj\" (UniqueName: \"kubernetes.io/projected/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-kube-api-access-bzwvj\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939710 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939718 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939726 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nx5h\" (UniqueName: \"kubernetes.io/projected/63fb1e51-5812-4adf-9e44-a7f4cd3349da-kube-api-access-8nx5h\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939746 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fb1e51-5812-4adf-9e44-a7f4cd3349da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:42 crc kubenswrapper[4838]: I1207 09:27:42.939754 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b4af2f-6fae-4a99-a9ee-8447c2a00df3-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.262944 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v59wv" event={"ID":"63fb1e51-5812-4adf-9e44-a7f4cd3349da","Type":"ContainerDied","Data":"e4485f84475dcdf96479dc1014e837b45f5fe783bd52c35aca3786e272ffc669"} Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.263282 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4485f84475dcdf96479dc1014e837b45f5fe783bd52c35aca3786e272ffc669" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.263353 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v59wv" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.280636 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" event={"ID":"99b4af2f-6fae-4a99-a9ee-8447c2a00df3","Type":"ContainerDied","Data":"6f3db32ad1f2f9d39c5febcda2399b34c1793c43c08200479a46888f0af95092"} Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.280680 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f3db32ad1f2f9d39c5febcda2399b34c1793c43c08200479a46888f0af95092" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.280750 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h7vd6" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.423481 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 07 09:27:43 crc kubenswrapper[4838]: E1207 09:27:43.423857 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b4af2f-6fae-4a99-a9ee-8447c2a00df3" containerName="nova-cell1-conductor-db-sync" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.423873 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b4af2f-6fae-4a99-a9ee-8447c2a00df3" containerName="nova-cell1-conductor-db-sync" Dec 07 09:27:43 crc kubenswrapper[4838]: E1207 09:27:43.423889 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fb1e51-5812-4adf-9e44-a7f4cd3349da" containerName="nova-manage" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.423896 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fb1e51-5812-4adf-9e44-a7f4cd3349da" containerName="nova-manage" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.424068 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="63fb1e51-5812-4adf-9e44-a7f4cd3349da" containerName="nova-manage" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.424094 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b4af2f-6fae-4a99-a9ee-8447c2a00df3" containerName="nova-cell1-conductor-db-sync" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.424629 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.431378 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.447697 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.552804 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.552923 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv9k5\" (UniqueName: \"kubernetes.io/projected/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-kube-api-access-sv9k5\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.552969 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.603173 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.603231 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.637667 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.637874 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-log" containerID="cri-o://48f87d65dc3b784f42b5961225c18c0feaa77fb4e6fc5c2c73b3651071e55b9a" gracePeriod=30 Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.638302 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-api" containerID="cri-o://996ff1647f93dc62377fe7f0c3b4d9f442b1a28f5e1d6642922ccf3f85c80916" gracePeriod=30 Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.648613 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.648799 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="bc296916-0117-4546-ae05-f9860fdf47b4" containerName="nova-scheduler-scheduler" containerID="cri-o://9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" gracePeriod=30 Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.658332 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.658431 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv9k5\" (UniqueName: \"kubernetes.io/projected/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-kube-api-access-sv9k5\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.658492 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.669567 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.669635 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.684439 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.716350 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv9k5\" (UniqueName: \"kubernetes.io/projected/8cdfa44f-0d15-4338-8a82-82b624b9ba6d-kube-api-access-sv9k5\") pod \"nova-cell1-conductor-0\" (UID: \"8cdfa44f-0d15-4338-8a82-82b624b9ba6d\") " pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:43 crc kubenswrapper[4838]: I1207 09:27:43.743340 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.239533 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.293280 4838 generic.go:334] "Generic (PLEG): container finished" podID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerID="48f87d65dc3b784f42b5961225c18c0feaa77fb4e6fc5c2c73b3651071e55b9a" exitCode=143 Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.293475 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8178ebbf-dd4e-4456-acfd-63ebbd26a95b","Type":"ContainerDied","Data":"48f87d65dc3b784f42b5961225c18c0feaa77fb4e6fc5c2c73b3651071e55b9a"} Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.295433 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-log" containerID="cri-o://4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436" gracePeriod=30 Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.295656 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8cdfa44f-0d15-4338-8a82-82b624b9ba6d","Type":"ContainerStarted","Data":"5488e746fe324e8e37f3dbfdba46dc9140cb8d110d34b72c081b67e838348b94"} Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.295947 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-metadata" containerID="cri-o://1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02" gracePeriod=30 Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.811565 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.881047 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-combined-ca-bundle\") pod \"72c05589-25ab-449e-b591-8865398f5431\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.881103 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d55jf\" (UniqueName: \"kubernetes.io/projected/72c05589-25ab-449e-b591-8865398f5431-kube-api-access-d55jf\") pod \"72c05589-25ab-449e-b591-8865398f5431\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.881191 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-nova-metadata-tls-certs\") pod \"72c05589-25ab-449e-b591-8865398f5431\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.881292 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c05589-25ab-449e-b591-8865398f5431-logs\") pod \"72c05589-25ab-449e-b591-8865398f5431\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.881895 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-config-data\") pod \"72c05589-25ab-449e-b591-8865398f5431\" (UID: \"72c05589-25ab-449e-b591-8865398f5431\") " Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.882003 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c05589-25ab-449e-b591-8865398f5431-logs" (OuterVolumeSpecName: "logs") pod "72c05589-25ab-449e-b591-8865398f5431" (UID: "72c05589-25ab-449e-b591-8865398f5431"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.882439 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c05589-25ab-449e-b591-8865398f5431-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.886867 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72c05589-25ab-449e-b591-8865398f5431-kube-api-access-d55jf" (OuterVolumeSpecName: "kube-api-access-d55jf") pod "72c05589-25ab-449e-b591-8865398f5431" (UID: "72c05589-25ab-449e-b591-8865398f5431"). InnerVolumeSpecName "kube-api-access-d55jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.921289 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72c05589-25ab-449e-b591-8865398f5431" (UID: "72c05589-25ab-449e-b591-8865398f5431"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.925006 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-config-data" (OuterVolumeSpecName: "config-data") pod "72c05589-25ab-449e-b591-8865398f5431" (UID: "72c05589-25ab-449e-b591-8865398f5431"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.939536 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "72c05589-25ab-449e-b591-8865398f5431" (UID: "72c05589-25ab-449e-b591-8865398f5431"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.984092 4838 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.984130 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.984141 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c05589-25ab-449e-b591-8865398f5431-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:44 crc kubenswrapper[4838]: I1207 09:27:44.984148 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d55jf\" (UniqueName: \"kubernetes.io/projected/72c05589-25ab-449e-b591-8865398f5431-kube-api-access-d55jf\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.306088 4838 generic.go:334] "Generic (PLEG): container finished" podID="72c05589-25ab-449e-b591-8865398f5431" containerID="1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02" exitCode=0 Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.306119 4838 generic.go:334] "Generic (PLEG): container finished" podID="72c05589-25ab-449e-b591-8865398f5431" containerID="4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436" exitCode=143 Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.306157 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"72c05589-25ab-449e-b591-8865398f5431","Type":"ContainerDied","Data":"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02"} Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.306184 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"72c05589-25ab-449e-b591-8865398f5431","Type":"ContainerDied","Data":"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436"} Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.306194 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"72c05589-25ab-449e-b591-8865398f5431","Type":"ContainerDied","Data":"e4839ddb9a9aacdea576243857b2a2ac10b4fd04db7925783195458fe18dce96"} Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.306209 4838 scope.go:117] "RemoveContainer" containerID="1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.306320 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.315722 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8cdfa44f-0d15-4338-8a82-82b624b9ba6d","Type":"ContainerStarted","Data":"a9b5b24da12c7c88c1b9e841edd91e1c0a841a29e2b93195c898fca2d13ccbfa"} Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.316064 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.341077 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.341050065 podStartE2EDuration="2.341050065s" podCreationTimestamp="2025-12-07 09:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:45.339385281 +0000 UTC m=+1282.046704298" watchObservedRunningTime="2025-12-07 09:27:45.341050065 +0000 UTC m=+1282.048369102" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.353415 4838 scope.go:117] "RemoveContainer" containerID="4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.373887 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.382165 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.385909 4838 scope.go:117] "RemoveContainer" containerID="1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02" Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.386338 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02\": container with ID starting with 1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02 not found: ID does not exist" containerID="1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.386382 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02"} err="failed to get container status \"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02\": rpc error: code = NotFound desc = could not find container \"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02\": container with ID starting with 1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02 not found: ID does not exist" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.386409 4838 scope.go:117] "RemoveContainer" containerID="4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436" Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.386801 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436\": container with ID starting with 4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436 not found: ID does not exist" containerID="4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.386835 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436"} err="failed to get container status \"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436\": rpc error: code = NotFound desc = could not find container \"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436\": container with ID starting with 4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436 not found: ID does not exist" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.386848 4838 scope.go:117] "RemoveContainer" containerID="1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.388012 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02"} err="failed to get container status \"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02\": rpc error: code = NotFound desc = could not find container \"1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02\": container with ID starting with 1402130b3a82bdd95f1d0aa0c0f5535e2e6c7ec04cac868263f8f13fc3626e02 not found: ID does not exist" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.388052 4838 scope.go:117] "RemoveContainer" containerID="4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.389020 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436"} err="failed to get container status \"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436\": rpc error: code = NotFound desc = could not find container \"4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436\": container with ID starting with 4dd6958cb33124d3e9166e957e4f64faec966fc4c2621e740f3b7b70b1e41436 not found: ID does not exist" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.394733 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.401203 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-metadata" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.401236 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-metadata" Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.401251 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-log" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.401258 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-log" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.401434 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-log" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.401446 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c05589-25ab-449e-b591-8865398f5431" containerName="nova-metadata-metadata" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.402365 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.404874 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.405935 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.408131 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.491375 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62drd\" (UniqueName: \"kubernetes.io/projected/d8a727b0-faa8-499a-bb85-c6234343f062-kube-api-access-62drd\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.491433 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-config-data\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.491481 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.491567 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a727b0-faa8-499a-bb85-c6234343f062-logs\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.491614 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.593604 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.593745 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a727b0-faa8-499a-bb85-c6234343f062-logs\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.593771 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.593829 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62drd\" (UniqueName: \"kubernetes.io/projected/d8a727b0-faa8-499a-bb85-c6234343f062-kube-api-access-62drd\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.593860 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-config-data\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.594172 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a727b0-faa8-499a-bb85-c6234343f062-logs\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.598432 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.598678 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-config-data\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.601887 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.615534 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62drd\" (UniqueName: \"kubernetes.io/projected/d8a727b0-faa8-499a-bb85-c6234343f062-kube-api-access-62drd\") pod \"nova-metadata-0\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.635011 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72c05589-25ab-449e-b591-8865398f5431" path="/var/lib/kubelet/pods/72c05589-25ab-449e-b591-8865398f5431/volumes" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.754905 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.805460 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.839487 4838 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.847277 4838 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.848464 4838 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 07 09:27:45 crc kubenswrapper[4838]: E1207 09:27:45.848521 4838 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="bc296916-0117-4546-ae05-f9860fdf47b4" containerName="nova-scheduler-scheduler" Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.880499 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-799db77f-dm9m6"] Dec 07 09:27:45 crc kubenswrapper[4838]: I1207 09:27:45.880782 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-799db77f-dm9m6" podUID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerName="dnsmasq-dns" containerID="cri-o://515d7287622568f214b4608ef14abf0c9ba0c7196d8311b875c520d5baa3b029" gracePeriod=10 Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.090912 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.341604 4838 generic.go:334] "Generic (PLEG): container finished" podID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerID="515d7287622568f214b4608ef14abf0c9ba0c7196d8311b875c520d5baa3b029" exitCode=0 Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.341908 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-dm9m6" event={"ID":"f64741e5-5b47-4f9c-bbeb-f0710259803a","Type":"ContainerDied","Data":"515d7287622568f214b4608ef14abf0c9ba0c7196d8311b875c520d5baa3b029"} Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.356378 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d8a727b0-faa8-499a-bb85-c6234343f062","Type":"ContainerStarted","Data":"f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c"} Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.356428 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d8a727b0-faa8-499a-bb85-c6234343f062","Type":"ContainerStarted","Data":"defebda0ffdcc66f198c1af6507bacf5e033e5bfa6a0f9d03dbc78068a5ab98f"} Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.408477 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.516448 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-dns-svc\") pod \"f64741e5-5b47-4f9c-bbeb-f0710259803a\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.516541 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smjj2\" (UniqueName: \"kubernetes.io/projected/f64741e5-5b47-4f9c-bbeb-f0710259803a-kube-api-access-smjj2\") pod \"f64741e5-5b47-4f9c-bbeb-f0710259803a\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.516582 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-sb\") pod \"f64741e5-5b47-4f9c-bbeb-f0710259803a\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.516674 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-config\") pod \"f64741e5-5b47-4f9c-bbeb-f0710259803a\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.516704 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-nb\") pod \"f64741e5-5b47-4f9c-bbeb-f0710259803a\" (UID: \"f64741e5-5b47-4f9c-bbeb-f0710259803a\") " Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.525470 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f64741e5-5b47-4f9c-bbeb-f0710259803a-kube-api-access-smjj2" (OuterVolumeSpecName: "kube-api-access-smjj2") pod "f64741e5-5b47-4f9c-bbeb-f0710259803a" (UID: "f64741e5-5b47-4f9c-bbeb-f0710259803a"). InnerVolumeSpecName "kube-api-access-smjj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.573626 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f64741e5-5b47-4f9c-bbeb-f0710259803a" (UID: "f64741e5-5b47-4f9c-bbeb-f0710259803a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.580846 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f64741e5-5b47-4f9c-bbeb-f0710259803a" (UID: "f64741e5-5b47-4f9c-bbeb-f0710259803a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.590171 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f64741e5-5b47-4f9c-bbeb-f0710259803a" (UID: "f64741e5-5b47-4f9c-bbeb-f0710259803a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.601627 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-config" (OuterVolumeSpecName: "config") pod "f64741e5-5b47-4f9c-bbeb-f0710259803a" (UID: "f64741e5-5b47-4f9c-bbeb-f0710259803a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.619278 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.619314 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smjj2\" (UniqueName: \"kubernetes.io/projected/f64741e5-5b47-4f9c-bbeb-f0710259803a-kube-api-access-smjj2\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.619330 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.619341 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:46 crc kubenswrapper[4838]: I1207 09:27:46.619350 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f64741e5-5b47-4f9c-bbeb-f0710259803a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.377594 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-dm9m6" event={"ID":"f64741e5-5b47-4f9c-bbeb-f0710259803a","Type":"ContainerDied","Data":"158cd92c67c9650653995fb3ce090c8d1c7242233e5d6869d7d5de1d7c3ae1c5"} Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.377872 4838 scope.go:117] "RemoveContainer" containerID="515d7287622568f214b4608ef14abf0c9ba0c7196d8311b875c520d5baa3b029" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.378007 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-dm9m6" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.386363 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d8a727b0-faa8-499a-bb85-c6234343f062","Type":"ContainerStarted","Data":"089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18"} Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.398291 4838 generic.go:334] "Generic (PLEG): container finished" podID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerID="996ff1647f93dc62377fe7f0c3b4d9f442b1a28f5e1d6642922ccf3f85c80916" exitCode=0 Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.399663 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8178ebbf-dd4e-4456-acfd-63ebbd26a95b","Type":"ContainerDied","Data":"996ff1647f93dc62377fe7f0c3b4d9f442b1a28f5e1d6642922ccf3f85c80916"} Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.407051 4838 scope.go:117] "RemoveContainer" containerID="08710852290d0b1e3569aa4b8249b62b6c08951238761f5a9b47f9ab11d29223" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.412777 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.412758969 podStartE2EDuration="2.412758969s" podCreationTimestamp="2025-12-07 09:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:47.409491553 +0000 UTC m=+1284.116810570" watchObservedRunningTime="2025-12-07 09:27:47.412758969 +0000 UTC m=+1284.120077986" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.450877 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-799db77f-dm9m6"] Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.467063 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-799db77f-dm9m6"] Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.543709 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.627532 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f64741e5-5b47-4f9c-bbeb-f0710259803a" path="/var/lib/kubelet/pods/f64741e5-5b47-4f9c-bbeb-f0710259803a/volumes" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.636423 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxt96\" (UniqueName: \"kubernetes.io/projected/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-kube-api-access-cxt96\") pod \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.636536 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-config-data\") pod \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.636579 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-combined-ca-bundle\") pod \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.636704 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-logs\") pod \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\" (UID: \"8178ebbf-dd4e-4456-acfd-63ebbd26a95b\") " Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.637369 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-logs" (OuterVolumeSpecName: "logs") pod "8178ebbf-dd4e-4456-acfd-63ebbd26a95b" (UID: "8178ebbf-dd4e-4456-acfd-63ebbd26a95b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.654597 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-kube-api-access-cxt96" (OuterVolumeSpecName: "kube-api-access-cxt96") pod "8178ebbf-dd4e-4456-acfd-63ebbd26a95b" (UID: "8178ebbf-dd4e-4456-acfd-63ebbd26a95b"). InnerVolumeSpecName "kube-api-access-cxt96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.664715 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8178ebbf-dd4e-4456-acfd-63ebbd26a95b" (UID: "8178ebbf-dd4e-4456-acfd-63ebbd26a95b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.674036 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-config-data" (OuterVolumeSpecName: "config-data") pod "8178ebbf-dd4e-4456-acfd-63ebbd26a95b" (UID: "8178ebbf-dd4e-4456-acfd-63ebbd26a95b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.740209 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxt96\" (UniqueName: \"kubernetes.io/projected/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-kube-api-access-cxt96\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.740243 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.740256 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:47 crc kubenswrapper[4838]: I1207 09:27:47.740264 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8178ebbf-dd4e-4456-acfd-63ebbd26a95b-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.310386 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.351323 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltrst\" (UniqueName: \"kubernetes.io/projected/bc296916-0117-4546-ae05-f9860fdf47b4-kube-api-access-ltrst\") pod \"bc296916-0117-4546-ae05-f9860fdf47b4\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.351361 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-combined-ca-bundle\") pod \"bc296916-0117-4546-ae05-f9860fdf47b4\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.351432 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-config-data\") pod \"bc296916-0117-4546-ae05-f9860fdf47b4\" (UID: \"bc296916-0117-4546-ae05-f9860fdf47b4\") " Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.356087 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc296916-0117-4546-ae05-f9860fdf47b4-kube-api-access-ltrst" (OuterVolumeSpecName: "kube-api-access-ltrst") pod "bc296916-0117-4546-ae05-f9860fdf47b4" (UID: "bc296916-0117-4546-ae05-f9860fdf47b4"). InnerVolumeSpecName "kube-api-access-ltrst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.384796 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc296916-0117-4546-ae05-f9860fdf47b4" (UID: "bc296916-0117-4546-ae05-f9860fdf47b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.384955 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-config-data" (OuterVolumeSpecName: "config-data") pod "bc296916-0117-4546-ae05-f9860fdf47b4" (UID: "bc296916-0117-4546-ae05-f9860fdf47b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.413849 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8178ebbf-dd4e-4456-acfd-63ebbd26a95b","Type":"ContainerDied","Data":"627b6fa374d4e24e25e61c607bfa4c7cfb954a3f84765d9e2992bb3e12f1f771"} Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.413915 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.413977 4838 scope.go:117] "RemoveContainer" containerID="996ff1647f93dc62377fe7f0c3b4d9f442b1a28f5e1d6642922ccf3f85c80916" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.417377 4838 generic.go:334] "Generic (PLEG): container finished" podID="bc296916-0117-4546-ae05-f9860fdf47b4" containerID="9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" exitCode=0 Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.417455 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bc296916-0117-4546-ae05-f9860fdf47b4","Type":"ContainerDied","Data":"9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19"} Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.417487 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bc296916-0117-4546-ae05-f9860fdf47b4","Type":"ContainerDied","Data":"a11b73967ca60c5e700fdcf981ae4153cc5e621810ae59b136b0085bc7f386cc"} Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.417548 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.446501 4838 scope.go:117] "RemoveContainer" containerID="48f87d65dc3b784f42b5961225c18c0feaa77fb4e6fc5c2c73b3651071e55b9a" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.465334 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltrst\" (UniqueName: \"kubernetes.io/projected/bc296916-0117-4546-ae05-f9860fdf47b4-kube-api-access-ltrst\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.465373 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.465386 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc296916-0117-4546-ae05-f9860fdf47b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.477060 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.543943 4838 scope.go:117] "RemoveContainer" containerID="9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.543967 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.554690 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568172 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: E1207 09:27:48.568582 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-log" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568599 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-log" Dec 07 09:27:48 crc kubenswrapper[4838]: E1207 09:27:48.568609 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerName="dnsmasq-dns" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568615 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerName="dnsmasq-dns" Dec 07 09:27:48 crc kubenswrapper[4838]: E1207 09:27:48.568627 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc296916-0117-4546-ae05-f9860fdf47b4" containerName="nova-scheduler-scheduler" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568633 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc296916-0117-4546-ae05-f9860fdf47b4" containerName="nova-scheduler-scheduler" Dec 07 09:27:48 crc kubenswrapper[4838]: E1207 09:27:48.568645 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-api" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568652 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-api" Dec 07 09:27:48 crc kubenswrapper[4838]: E1207 09:27:48.568661 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerName="init" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568666 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerName="init" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568851 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-api" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568867 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" containerName="nova-api-log" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568877 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc296916-0117-4546-ae05-f9860fdf47b4" containerName="nova-scheduler-scheduler" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.568893 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f64741e5-5b47-4f9c-bbeb-f0710259803a" containerName="dnsmasq-dns" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.569481 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.578328 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.579323 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.610786 4838 scope.go:117] "RemoveContainer" containerID="9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" Dec 07 09:27:48 crc kubenswrapper[4838]: E1207 09:27:48.611271 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19\": container with ID starting with 9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19 not found: ID does not exist" containerID="9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.611365 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19"} err="failed to get container status \"9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19\": rpc error: code = NotFound desc = could not find container \"9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19\": container with ID starting with 9739d944a3ec910de738bb24d690f4a37137c943e591765792225e5514b58b19 not found: ID does not exist" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.623952 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.631937 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.633478 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.635382 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.639394 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.671850 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drhh6\" (UniqueName: \"kubernetes.io/projected/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-kube-api-access-drhh6\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.672004 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.672062 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-config-data\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.773564 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.773645 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdt4j\" (UniqueName: \"kubernetes.io/projected/7c385659-08ed-4099-b753-7287100bda19-kube-api-access-kdt4j\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.773690 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.773717 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-config-data\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.773745 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-config-data\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.774021 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c385659-08ed-4099-b753-7287100bda19-logs\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.774184 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drhh6\" (UniqueName: \"kubernetes.io/projected/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-kube-api-access-drhh6\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.780941 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.780986 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-config-data\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.789207 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drhh6\" (UniqueName: \"kubernetes.io/projected/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-kube-api-access-drhh6\") pod \"nova-scheduler-0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.876795 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c385659-08ed-4099-b753-7287100bda19-logs\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.876903 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.876938 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdt4j\" (UniqueName: \"kubernetes.io/projected/7c385659-08ed-4099-b753-7287100bda19-kube-api-access-kdt4j\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.876971 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-config-data\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.877389 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c385659-08ed-4099-b753-7287100bda19-logs\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.880341 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.882515 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-config-data\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.904753 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdt4j\" (UniqueName: \"kubernetes.io/projected/7c385659-08ed-4099-b753-7287100bda19-kube-api-access-kdt4j\") pod \"nova-api-0\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " pod="openstack/nova-api-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.907111 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:27:48 crc kubenswrapper[4838]: I1207 09:27:48.987398 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:27:49 crc kubenswrapper[4838]: W1207 09:27:49.461338 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d27a2f0_b17d_4d7b_9342_e5b5e22b8da0.slice/crio-c307aafc0971e53d85ced2f15b30bda49ea0457ae40428ede38eba57e6320cb1 WatchSource:0}: Error finding container c307aafc0971e53d85ced2f15b30bda49ea0457ae40428ede38eba57e6320cb1: Status 404 returned error can't find the container with id c307aafc0971e53d85ced2f15b30bda49ea0457ae40428ede38eba57e6320cb1 Dec 07 09:27:49 crc kubenswrapper[4838]: I1207 09:27:49.463751 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:27:49 crc kubenswrapper[4838]: W1207 09:27:49.561440 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c385659_08ed_4099_b753_7287100bda19.slice/crio-964f11a65bbc8ed107b1eaa92b03066bbd094c8bae5fe2b1cd93b25082846796 WatchSource:0}: Error finding container 964f11a65bbc8ed107b1eaa92b03066bbd094c8bae5fe2b1cd93b25082846796: Status 404 returned error can't find the container with id 964f11a65bbc8ed107b1eaa92b03066bbd094c8bae5fe2b1cd93b25082846796 Dec 07 09:27:49 crc kubenswrapper[4838]: I1207 09:27:49.562091 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:27:49 crc kubenswrapper[4838]: I1207 09:27:49.629725 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8178ebbf-dd4e-4456-acfd-63ebbd26a95b" path="/var/lib/kubelet/pods/8178ebbf-dd4e-4456-acfd-63ebbd26a95b/volumes" Dec 07 09:27:49 crc kubenswrapper[4838]: I1207 09:27:49.630596 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc296916-0117-4546-ae05-f9860fdf47b4" path="/var/lib/kubelet/pods/bc296916-0117-4546-ae05-f9860fdf47b4/volumes" Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.445030 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c385659-08ed-4099-b753-7287100bda19","Type":"ContainerStarted","Data":"3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad"} Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.445630 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c385659-08ed-4099-b753-7287100bda19","Type":"ContainerStarted","Data":"28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025"} Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.445789 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c385659-08ed-4099-b753-7287100bda19","Type":"ContainerStarted","Data":"964f11a65bbc8ed107b1eaa92b03066bbd094c8bae5fe2b1cd93b25082846796"} Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.448412 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0","Type":"ContainerStarted","Data":"22b43f0b83070e3c6e92d8f2c0e7f80d9f9dbe3dbcd9f39a068578d945a048ec"} Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.448556 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0","Type":"ContainerStarted","Data":"c307aafc0971e53d85ced2f15b30bda49ea0457ae40428ede38eba57e6320cb1"} Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.468453 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.468434182 podStartE2EDuration="2.468434182s" podCreationTimestamp="2025-12-07 09:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:50.461363456 +0000 UTC m=+1287.168682493" watchObservedRunningTime="2025-12-07 09:27:50.468434182 +0000 UTC m=+1287.175753199" Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.494446 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.494421166 podStartE2EDuration="2.494421166s" podCreationTimestamp="2025-12-07 09:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:27:50.47785711 +0000 UTC m=+1287.185176147" watchObservedRunningTime="2025-12-07 09:27:50.494421166 +0000 UTC m=+1287.201740183" Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.755797 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 09:27:50 crc kubenswrapper[4838]: I1207 09:27:50.756065 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 09:27:53 crc kubenswrapper[4838]: I1207 09:27:53.768644 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 07 09:27:53 crc kubenswrapper[4838]: I1207 09:27:53.907345 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 07 09:27:54 crc kubenswrapper[4838]: I1207 09:27:54.492683 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:27:54 crc kubenswrapper[4838]: I1207 09:27:54.493042 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:27:55 crc kubenswrapper[4838]: I1207 09:27:55.756031 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 09:27:55 crc kubenswrapper[4838]: I1207 09:27:55.756084 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 09:27:56 crc kubenswrapper[4838]: I1207 09:27:56.767032 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.173:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 09:27:56 crc kubenswrapper[4838]: I1207 09:27:56.767064 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.173:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 09:27:58 crc kubenswrapper[4838]: I1207 09:27:58.907569 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 07 09:27:58 crc kubenswrapper[4838]: I1207 09:27:58.942652 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 07 09:27:58 crc kubenswrapper[4838]: I1207 09:27:58.988463 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:27:58 crc kubenswrapper[4838]: I1207 09:27:58.988520 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:27:59 crc kubenswrapper[4838]: I1207 09:27:59.551590 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 07 09:28:00 crc kubenswrapper[4838]: I1207 09:28:00.070043 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.175:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:00 crc kubenswrapper[4838]: I1207 09:28:00.070422 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.175:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:05 crc kubenswrapper[4838]: I1207 09:28:05.760772 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 09:28:05 crc kubenswrapper[4838]: I1207 09:28:05.761397 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 09:28:05 crc kubenswrapper[4838]: I1207 09:28:05.768420 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 09:28:05 crc kubenswrapper[4838]: I1207 09:28:05.768792 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.578574 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.587272 4838 generic.go:334] "Generic (PLEG): container finished" podID="81adc2d0-f8c4-4b96-b459-9a446d468111" containerID="29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf" exitCode=137 Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.587901 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"81adc2d0-f8c4-4b96-b459-9a446d468111","Type":"ContainerDied","Data":"29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf"} Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.587946 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"81adc2d0-f8c4-4b96-b459-9a446d468111","Type":"ContainerDied","Data":"32b02e1c454cc3a9dcdba84cae6ac2d9e97eebd72dec405fcf1d3dafe54ea213"} Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.587966 4838 scope.go:117] "RemoveContainer" containerID="29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.588099 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.627717 4838 scope.go:117] "RemoveContainer" containerID="29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf" Dec 07 09:28:06 crc kubenswrapper[4838]: E1207 09:28:06.628466 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf\": container with ID starting with 29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf not found: ID does not exist" containerID="29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.628512 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf"} err="failed to get container status \"29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf\": rpc error: code = NotFound desc = could not find container \"29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf\": container with ID starting with 29dc2abefd33bb92d2c782d7475a806eccb6f5407241ddaaeb0aa4d6fae73ccf not found: ID does not exist" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.703361 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncwh8\" (UniqueName: \"kubernetes.io/projected/81adc2d0-f8c4-4b96-b459-9a446d468111-kube-api-access-ncwh8\") pod \"81adc2d0-f8c4-4b96-b459-9a446d468111\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.703542 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-combined-ca-bundle\") pod \"81adc2d0-f8c4-4b96-b459-9a446d468111\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.703610 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-config-data\") pod \"81adc2d0-f8c4-4b96-b459-9a446d468111\" (UID: \"81adc2d0-f8c4-4b96-b459-9a446d468111\") " Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.722840 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81adc2d0-f8c4-4b96-b459-9a446d468111-kube-api-access-ncwh8" (OuterVolumeSpecName: "kube-api-access-ncwh8") pod "81adc2d0-f8c4-4b96-b459-9a446d468111" (UID: "81adc2d0-f8c4-4b96-b459-9a446d468111"). InnerVolumeSpecName "kube-api-access-ncwh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.730781 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81adc2d0-f8c4-4b96-b459-9a446d468111" (UID: "81adc2d0-f8c4-4b96-b459-9a446d468111"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.732348 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-config-data" (OuterVolumeSpecName: "config-data") pod "81adc2d0-f8c4-4b96-b459-9a446d468111" (UID: "81adc2d0-f8c4-4b96-b459-9a446d468111"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.806042 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.806078 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncwh8\" (UniqueName: \"kubernetes.io/projected/81adc2d0-f8c4-4b96-b459-9a446d468111-kube-api-access-ncwh8\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.806092 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81adc2d0-f8c4-4b96-b459-9a446d468111-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.938319 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.945627 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.960172 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:28:06 crc kubenswrapper[4838]: E1207 09:28:06.960527 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81adc2d0-f8c4-4b96-b459-9a446d468111" containerName="nova-cell1-novncproxy-novncproxy" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.960541 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="81adc2d0-f8c4-4b96-b459-9a446d468111" containerName="nova-cell1-novncproxy-novncproxy" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.960711 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="81adc2d0-f8c4-4b96-b459-9a446d468111" containerName="nova-cell1-novncproxy-novncproxy" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.961298 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.964649 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.965003 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.965240 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 07 09:28:06 crc kubenswrapper[4838]: I1207 09:28:06.981104 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.009479 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4thq\" (UniqueName: \"kubernetes.io/projected/cb3ea478-6a63-4e03-b9c1-890fca36b984-kube-api-access-x4thq\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.009607 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.009644 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.009692 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.009723 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.110952 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.110999 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.111058 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4thq\" (UniqueName: \"kubernetes.io/projected/cb3ea478-6a63-4e03-b9c1-890fca36b984-kube-api-access-x4thq\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.111132 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.111158 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.115944 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.116130 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.117504 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.117992 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb3ea478-6a63-4e03-b9c1-890fca36b984-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.127090 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4thq\" (UniqueName: \"kubernetes.io/projected/cb3ea478-6a63-4e03-b9c1-890fca36b984-kube-api-access-x4thq\") pod \"nova-cell1-novncproxy-0\" (UID: \"cb3ea478-6a63-4e03-b9c1-890fca36b984\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.279580 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.628345 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81adc2d0-f8c4-4b96-b459-9a446d468111" path="/var/lib/kubelet/pods/81adc2d0-f8c4-4b96-b459-9a446d468111/volumes" Dec 07 09:28:07 crc kubenswrapper[4838]: I1207 09:28:07.779787 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 09:28:08 crc kubenswrapper[4838]: I1207 09:28:08.621020 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb3ea478-6a63-4e03-b9c1-890fca36b984","Type":"ContainerStarted","Data":"6a933ecede031fc834e6771f178607290443bc61457e5254e3d080dddfdc7abb"} Dec 07 09:28:08 crc kubenswrapper[4838]: I1207 09:28:08.621383 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cb3ea478-6a63-4e03-b9c1-890fca36b984","Type":"ContainerStarted","Data":"c891a99e534c229b1008d7a6c63488bb3cc5764b843ae1bdbbbfb75f0f897430"} Dec 07 09:28:08 crc kubenswrapper[4838]: I1207 09:28:08.993551 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 09:28:08 crc kubenswrapper[4838]: I1207 09:28:08.993610 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 09:28:08 crc kubenswrapper[4838]: I1207 09:28:08.994576 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 09:28:08 crc kubenswrapper[4838]: I1207 09:28:08.994761 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.000116 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.001337 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.018405 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.018389353 podStartE2EDuration="3.018389353s" podCreationTimestamp="2025-12-07 09:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:28:08.640537104 +0000 UTC m=+1305.347856131" watchObservedRunningTime="2025-12-07 09:28:09.018389353 +0000 UTC m=+1305.725708370" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.215331 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f64966d5-lsr6s"] Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.217543 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.226655 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-lsr6s"] Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.256407 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-nb\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.256503 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgk4g\" (UniqueName: \"kubernetes.io/projected/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-kube-api-access-rgk4g\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.256528 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-config\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.256573 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-sb\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.256684 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-dns-svc\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.358353 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-nb\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.358432 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgk4g\" (UniqueName: \"kubernetes.io/projected/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-kube-api-access-rgk4g\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.358466 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-config\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.358496 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-sb\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.358569 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-dns-svc\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.359742 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-nb\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.359857 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-dns-svc\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.360597 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-sb\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.360734 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-config\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.384188 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgk4g\" (UniqueName: \"kubernetes.io/projected/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-kube-api-access-rgk4g\") pod \"dnsmasq-dns-f64966d5-lsr6s\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:09 crc kubenswrapper[4838]: I1207 09:28:09.540130 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:10 crc kubenswrapper[4838]: W1207 09:28:10.158111 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45c647ef_5d0b_44ea_96f9_8c4c00c237f9.slice/crio-c83338f81916872e66027b04df0bada58447605a1f124fcee72bd5585ddc0830 WatchSource:0}: Error finding container c83338f81916872e66027b04df0bada58447605a1f124fcee72bd5585ddc0830: Status 404 returned error can't find the container with id c83338f81916872e66027b04df0bada58447605a1f124fcee72bd5585ddc0830 Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.164051 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-lsr6s"] Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.645578 4838 generic.go:334] "Generic (PLEG): container finished" podID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerID="157f46c3097105f4976e547241583cf8932c11f8a1ff77a88bd7deeaf723e27b" exitCode=0 Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.645691 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" event={"ID":"45c647ef-5d0b-44ea-96f9-8c4c00c237f9","Type":"ContainerDied","Data":"157f46c3097105f4976e547241583cf8932c11f8a1ff77a88bd7deeaf723e27b"} Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.646024 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" event={"ID":"45c647ef-5d0b-44ea-96f9-8c4c00c237f9","Type":"ContainerStarted","Data":"c83338f81916872e66027b04df0bada58447605a1f124fcee72bd5585ddc0830"} Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.874087 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.874346 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-central-agent" containerID="cri-o://087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6" gracePeriod=30 Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.874524 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-notification-agent" containerID="cri-o://ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924" gracePeriod=30 Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.874534 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="sg-core" containerID="cri-o://0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d" gracePeriod=30 Dec 07 09:28:10 crc kubenswrapper[4838]: I1207 09:28:10.877448 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="proxy-httpd" containerID="cri-o://eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806" gracePeriod=30 Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.435222 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.162:3000/\": dial tcp 10.217.0.162:3000: connect: connection refused" Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.664370 4838 generic.go:334] "Generic (PLEG): container finished" podID="693d763b-81c2-4279-9160-668a1819e01b" containerID="eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806" exitCode=0 Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.664406 4838 generic.go:334] "Generic (PLEG): container finished" podID="693d763b-81c2-4279-9160-668a1819e01b" containerID="0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d" exitCode=2 Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.664417 4838 generic.go:334] "Generic (PLEG): container finished" podID="693d763b-81c2-4279-9160-668a1819e01b" containerID="087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6" exitCode=0 Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.664461 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerDied","Data":"eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806"} Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.664490 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerDied","Data":"0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d"} Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.664503 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerDied","Data":"087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6"} Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.666885 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" event={"ID":"45c647ef-5d0b-44ea-96f9-8c4c00c237f9","Type":"ContainerStarted","Data":"4dc72005e29e40e16a332f0c82e3ea5dec6046c64ef78695f7f8b1cda5882398"} Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.667061 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.696593 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" podStartSLOduration=2.696574509 podStartE2EDuration="2.696574509s" podCreationTimestamp="2025-12-07 09:28:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:28:11.69243681 +0000 UTC m=+1308.399755867" watchObservedRunningTime="2025-12-07 09:28:11.696574509 +0000 UTC m=+1308.403893536" Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.780688 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.780967 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-log" containerID="cri-o://28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025" gracePeriod=30 Dec 07 09:28:11 crc kubenswrapper[4838]: I1207 09:28:11.781136 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-api" containerID="cri-o://3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad" gracePeriod=30 Dec 07 09:28:12 crc kubenswrapper[4838]: I1207 09:28:12.280570 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:12 crc kubenswrapper[4838]: I1207 09:28:12.674972 4838 generic.go:334] "Generic (PLEG): container finished" podID="7c385659-08ed-4099-b753-7287100bda19" containerID="28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025" exitCode=143 Dec 07 09:28:12 crc kubenswrapper[4838]: I1207 09:28:12.675735 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c385659-08ed-4099-b753-7287100bda19","Type":"ContainerDied","Data":"28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025"} Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.392942 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.511394 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c385659-08ed-4099-b753-7287100bda19-logs\") pod \"7c385659-08ed-4099-b753-7287100bda19\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.511529 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdt4j\" (UniqueName: \"kubernetes.io/projected/7c385659-08ed-4099-b753-7287100bda19-kube-api-access-kdt4j\") pod \"7c385659-08ed-4099-b753-7287100bda19\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.511582 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-combined-ca-bundle\") pod \"7c385659-08ed-4099-b753-7287100bda19\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.511677 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-config-data\") pod \"7c385659-08ed-4099-b753-7287100bda19\" (UID: \"7c385659-08ed-4099-b753-7287100bda19\") " Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.511957 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c385659-08ed-4099-b753-7287100bda19-logs" (OuterVolumeSpecName: "logs") pod "7c385659-08ed-4099-b753-7287100bda19" (UID: "7c385659-08ed-4099-b753-7287100bda19"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.512149 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c385659-08ed-4099-b753-7287100bda19-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.533183 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c385659-08ed-4099-b753-7287100bda19-kube-api-access-kdt4j" (OuterVolumeSpecName: "kube-api-access-kdt4j") pod "7c385659-08ed-4099-b753-7287100bda19" (UID: "7c385659-08ed-4099-b753-7287100bda19"). InnerVolumeSpecName "kube-api-access-kdt4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.543667 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-config-data" (OuterVolumeSpecName: "config-data") pod "7c385659-08ed-4099-b753-7287100bda19" (UID: "7c385659-08ed-4099-b753-7287100bda19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.551413 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c385659-08ed-4099-b753-7287100bda19" (UID: "7c385659-08ed-4099-b753-7287100bda19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.613783 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.613895 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdt4j\" (UniqueName: \"kubernetes.io/projected/7c385659-08ed-4099-b753-7287100bda19-kube-api-access-kdt4j\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.613912 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c385659-08ed-4099-b753-7287100bda19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.700617 4838 generic.go:334] "Generic (PLEG): container finished" podID="7c385659-08ed-4099-b753-7287100bda19" containerID="3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad" exitCode=0 Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.700667 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c385659-08ed-4099-b753-7287100bda19","Type":"ContainerDied","Data":"3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad"} Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.700697 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c385659-08ed-4099-b753-7287100bda19","Type":"ContainerDied","Data":"964f11a65bbc8ed107b1eaa92b03066bbd094c8bae5fe2b1cd93b25082846796"} Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.700667 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.700721 4838 scope.go:117] "RemoveContainer" containerID="3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.722769 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.727042 4838 scope.go:117] "RemoveContainer" containerID="28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.734823 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.747444 4838 scope.go:117] "RemoveContainer" containerID="3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad" Dec 07 09:28:15 crc kubenswrapper[4838]: E1207 09:28:15.748432 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad\": container with ID starting with 3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad not found: ID does not exist" containerID="3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.748536 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad"} err="failed to get container status \"3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad\": rpc error: code = NotFound desc = could not find container \"3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad\": container with ID starting with 3a63547e9e5e7cc61aa4969e5a41d797ea60bda51b18b6bff1bf17b75bf96cad not found: ID does not exist" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.748671 4838 scope.go:117] "RemoveContainer" containerID="28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025" Dec 07 09:28:15 crc kubenswrapper[4838]: E1207 09:28:15.749994 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025\": container with ID starting with 28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025 not found: ID does not exist" containerID="28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.750100 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025"} err="failed to get container status \"28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025\": rpc error: code = NotFound desc = could not find container \"28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025\": container with ID starting with 28227c4e7acc452928c8b48454854eb17a806495916fd4a29c63ff170a787025 not found: ID does not exist" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.754652 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:15 crc kubenswrapper[4838]: E1207 09:28:15.755761 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-log" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.755889 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-log" Dec 07 09:28:15 crc kubenswrapper[4838]: E1207 09:28:15.756049 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-api" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.756123 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-api" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.756349 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-api" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.756428 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c385659-08ed-4099-b753-7287100bda19" containerName="nova-api-log" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.757565 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.759502 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.759547 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.761020 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.767471 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.919606 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-config-data\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.919906 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ca3256-1182-42cc-862f-58d99bec1b42-logs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.920065 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-public-tls-certs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.920187 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.920293 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:15 crc kubenswrapper[4838]: I1207 09:28:15.920479 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlvvv\" (UniqueName: \"kubernetes.io/projected/30ca3256-1182-42cc-862f-58d99bec1b42-kube-api-access-zlvvv\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.023443 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-config-data\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.023503 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ca3256-1182-42cc-862f-58d99bec1b42-logs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.023533 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-public-tls-certs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.023551 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.023566 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.023593 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlvvv\" (UniqueName: \"kubernetes.io/projected/30ca3256-1182-42cc-862f-58d99bec1b42-kube-api-access-zlvvv\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.025856 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ca3256-1182-42cc-862f-58d99bec1b42-logs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.030675 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-config-data\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.041325 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-public-tls-certs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.041840 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.042457 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.047803 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlvvv\" (UniqueName: \"kubernetes.io/projected/30ca3256-1182-42cc-862f-58d99bec1b42-kube-api-access-zlvvv\") pod \"nova-api-0\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.074452 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.248359 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430394 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-config-data\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430452 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-sg-core-conf-yaml\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430487 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-ceilometer-tls-certs\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430527 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-log-httpd\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430595 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-run-httpd\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430624 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwfsz\" (UniqueName: \"kubernetes.io/projected/693d763b-81c2-4279-9160-668a1819e01b-kube-api-access-xwfsz\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430735 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-scripts\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.430767 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-combined-ca-bundle\") pod \"693d763b-81c2-4279-9160-668a1819e01b\" (UID: \"693d763b-81c2-4279-9160-668a1819e01b\") " Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.432237 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.432671 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.439533 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693d763b-81c2-4279-9160-668a1819e01b-kube-api-access-xwfsz" (OuterVolumeSpecName: "kube-api-access-xwfsz") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "kube-api-access-xwfsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.439638 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-scripts" (OuterVolumeSpecName: "scripts") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.477297 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.502864 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.526984 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.533075 4838 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.533103 4838 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.533114 4838 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.533122 4838 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/693d763b-81c2-4279-9160-668a1819e01b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.533130 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwfsz\" (UniqueName: \"kubernetes.io/projected/693d763b-81c2-4279-9160-668a1819e01b-kube-api-access-xwfsz\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.533151 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.533159 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.545169 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-config-data" (OuterVolumeSpecName: "config-data") pod "693d763b-81c2-4279-9160-668a1819e01b" (UID: "693d763b-81c2-4279-9160-668a1819e01b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.605089 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.635228 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693d763b-81c2-4279-9160-668a1819e01b-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.716681 4838 generic.go:334] "Generic (PLEG): container finished" podID="693d763b-81c2-4279-9160-668a1819e01b" containerID="ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924" exitCode=0 Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.716737 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.716762 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerDied","Data":"ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924"} Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.716792 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"693d763b-81c2-4279-9160-668a1819e01b","Type":"ContainerDied","Data":"739113d5b7a9057093649e9e0cec4617886259f096d663e62cb127c39e29bdd7"} Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.716809 4838 scope.go:117] "RemoveContainer" containerID="eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.721803 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30ca3256-1182-42cc-862f-58d99bec1b42","Type":"ContainerStarted","Data":"ba464b85a3a20d697683de69b8fded68b4491d17b30c75a98897136ec46a1ced"} Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.743548 4838 scope.go:117] "RemoveContainer" containerID="0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.774030 4838 scope.go:117] "RemoveContainer" containerID="ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.774517 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.794020 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.804919 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.805360 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-notification-agent" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805384 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-notification-agent" Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.805403 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="sg-core" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805412 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="sg-core" Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.805424 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="proxy-httpd" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805431 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="proxy-httpd" Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.805458 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-central-agent" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805468 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-central-agent" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805665 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-notification-agent" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805681 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="ceilometer-central-agent" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805699 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="sg-core" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.805713 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="693d763b-81c2-4279-9160-668a1819e01b" containerName="proxy-httpd" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.807828 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.811028 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.811176 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.811318 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.817774 4838 scope.go:117] "RemoveContainer" containerID="087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.819544 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.859038 4838 scope.go:117] "RemoveContainer" containerID="eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806" Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.859876 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806\": container with ID starting with eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806 not found: ID does not exist" containerID="eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.859902 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806"} err="failed to get container status \"eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806\": rpc error: code = NotFound desc = could not find container \"eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806\": container with ID starting with eea57651c75e82436cca76ff38d22a625fb66287374c3f1e579dc604f9c6a806 not found: ID does not exist" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.859920 4838 scope.go:117] "RemoveContainer" containerID="0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d" Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.860113 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d\": container with ID starting with 0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d not found: ID does not exist" containerID="0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.860129 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d"} err="failed to get container status \"0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d\": rpc error: code = NotFound desc = could not find container \"0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d\": container with ID starting with 0a00b68ef2a91059ebf71c4c0841ad942918d716969f148bc55b3574b895be3d not found: ID does not exist" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.860156 4838 scope.go:117] "RemoveContainer" containerID="ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924" Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.860333 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924\": container with ID starting with ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924 not found: ID does not exist" containerID="ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.860351 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924"} err="failed to get container status \"ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924\": rpc error: code = NotFound desc = could not find container \"ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924\": container with ID starting with ee4f4e474a722eadb43874cdabc684da3b9cf709c56a05d3ead6fa1b43959924 not found: ID does not exist" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.860363 4838 scope.go:117] "RemoveContainer" containerID="087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6" Dec 07 09:28:16 crc kubenswrapper[4838]: E1207 09:28:16.860528 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6\": container with ID starting with 087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6 not found: ID does not exist" containerID="087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.860542 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6"} err="failed to get container status \"087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6\": rpc error: code = NotFound desc = could not find container \"087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6\": container with ID starting with 087aa4b0c71fa3e09b00e8d832391ed89eeb4f5397cbd5ec1e85e383a28105f6 not found: ID does not exist" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941029 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941076 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941102 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-run-httpd\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941119 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-log-httpd\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941296 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-config-data\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941499 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc9zq\" (UniqueName: \"kubernetes.io/projected/d05b7eba-6424-44d0-bcf9-822d5305f798-kube-api-access-zc9zq\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941612 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:16 crc kubenswrapper[4838]: I1207 09:28:16.941667 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-scripts\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043305 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-config-data\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043393 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc9zq\" (UniqueName: \"kubernetes.io/projected/d05b7eba-6424-44d0-bcf9-822d5305f798-kube-api-access-zc9zq\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043438 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043465 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-scripts\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043502 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043524 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043546 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-run-httpd\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043562 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-log-httpd\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.043976 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-log-httpd\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.045236 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-run-httpd\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.048851 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-config-data\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.050978 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-scripts\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.051600 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.052463 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.053167 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.068578 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc9zq\" (UniqueName: \"kubernetes.io/projected/d05b7eba-6424-44d0-bcf9-822d5305f798-kube-api-access-zc9zq\") pod \"ceilometer-0\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.163587 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.280503 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.302139 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.630896 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="693d763b-81c2-4279-9160-668a1819e01b" path="/var/lib/kubelet/pods/693d763b-81c2-4279-9160-668a1819e01b/volumes" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.632319 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c385659-08ed-4099-b753-7287100bda19" path="/var/lib/kubelet/pods/7c385659-08ed-4099-b753-7287100bda19/volumes" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.632968 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.633018 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.733467 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30ca3256-1182-42cc-862f-58d99bec1b42","Type":"ContainerStarted","Data":"95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8"} Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.733534 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30ca3256-1182-42cc-862f-58d99bec1b42","Type":"ContainerStarted","Data":"02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5"} Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.735033 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerStarted","Data":"6a47691762c3038cbb5f49f0c888d76035bacb036fc3dc1174ea94d6c2ab847b"} Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.758186 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.758166343 podStartE2EDuration="2.758166343s" podCreationTimestamp="2025-12-07 09:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:28:17.751874707 +0000 UTC m=+1314.459193734" watchObservedRunningTime="2025-12-07 09:28:17.758166343 +0000 UTC m=+1314.465485360" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.765534 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.951074 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-wndph"] Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.954646 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.969250 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.969571 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 07 09:28:17 crc kubenswrapper[4838]: I1207 09:28:17.977883 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-wndph"] Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.060909 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2nbx\" (UniqueName: \"kubernetes.io/projected/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-kube-api-access-f2nbx\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.061137 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-config-data\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.061287 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-scripts\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.061453 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.163695 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.163886 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2nbx\" (UniqueName: \"kubernetes.io/projected/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-kube-api-access-f2nbx\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.163969 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-config-data\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.164096 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-scripts\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.168996 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-scripts\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.170303 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.170517 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-config-data\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.180429 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2nbx\" (UniqueName: \"kubernetes.io/projected/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-kube-api-access-f2nbx\") pod \"nova-cell1-cell-mapping-wndph\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.376048 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.747188 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerStarted","Data":"fe6fe567f47b9d35a5f388d8c252e82dc50c4a97b23a78cf6bfae9612098f1b1"} Dec 07 09:28:18 crc kubenswrapper[4838]: W1207 09:28:18.803955 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9fd40d9_0ac7_47d6_afd7_05a6a88ef137.slice/crio-b8b2878ae0f6117d792fa811d8536920cdcbca3c360be4e8674e1536982ec952 WatchSource:0}: Error finding container b8b2878ae0f6117d792fa811d8536920cdcbca3c360be4e8674e1536982ec952: Status 404 returned error can't find the container with id b8b2878ae0f6117d792fa811d8536920cdcbca3c360be4e8674e1536982ec952 Dec 07 09:28:18 crc kubenswrapper[4838]: I1207 09:28:18.806215 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-wndph"] Dec 07 09:28:19 crc kubenswrapper[4838]: I1207 09:28:19.542051 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:28:19 crc kubenswrapper[4838]: I1207 09:28:19.634673 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-7f9tn"] Dec 07 09:28:19 crc kubenswrapper[4838]: I1207 09:28:19.634954 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" podUID="81515be6-4b04-4629-9ff8-37362ab0a313" containerName="dnsmasq-dns" containerID="cri-o://3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2" gracePeriod=10 Dec 07 09:28:19 crc kubenswrapper[4838]: I1207 09:28:19.802771 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerStarted","Data":"b1eb11aac662b53da64cff6ccda2fc89c6fc85dcedc877014663a7c2491bc73f"} Dec 07 09:28:19 crc kubenswrapper[4838]: I1207 09:28:19.813440 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wndph" event={"ID":"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137","Type":"ContainerStarted","Data":"b25b83605bcbe9e2002ca0fe388109c280cc08dd827f6dc0b00c740c1151d4ca"} Dec 07 09:28:19 crc kubenswrapper[4838]: I1207 09:28:19.813478 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wndph" event={"ID":"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137","Type":"ContainerStarted","Data":"b8b2878ae0f6117d792fa811d8536920cdcbca3c360be4e8674e1536982ec952"} Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.284281 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.302398 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-wndph" podStartSLOduration=3.302379248 podStartE2EDuration="3.302379248s" podCreationTimestamp="2025-12-07 09:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:28:19.836151717 +0000 UTC m=+1316.543470734" watchObservedRunningTime="2025-12-07 09:28:20.302379248 +0000 UTC m=+1317.009698265" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.415981 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-config\") pod \"81515be6-4b04-4629-9ff8-37362ab0a313\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.416092 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct6mw\" (UniqueName: \"kubernetes.io/projected/81515be6-4b04-4629-9ff8-37362ab0a313-kube-api-access-ct6mw\") pod \"81515be6-4b04-4629-9ff8-37362ab0a313\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.416178 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-dns-svc\") pod \"81515be6-4b04-4629-9ff8-37362ab0a313\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.416219 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-sb\") pod \"81515be6-4b04-4629-9ff8-37362ab0a313\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.416239 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-nb\") pod \"81515be6-4b04-4629-9ff8-37362ab0a313\" (UID: \"81515be6-4b04-4629-9ff8-37362ab0a313\") " Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.421650 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81515be6-4b04-4629-9ff8-37362ab0a313-kube-api-access-ct6mw" (OuterVolumeSpecName: "kube-api-access-ct6mw") pod "81515be6-4b04-4629-9ff8-37362ab0a313" (UID: "81515be6-4b04-4629-9ff8-37362ab0a313"). InnerVolumeSpecName "kube-api-access-ct6mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.475338 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-config" (OuterVolumeSpecName: "config") pod "81515be6-4b04-4629-9ff8-37362ab0a313" (UID: "81515be6-4b04-4629-9ff8-37362ab0a313"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.485425 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81515be6-4b04-4629-9ff8-37362ab0a313" (UID: "81515be6-4b04-4629-9ff8-37362ab0a313"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.488329 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "81515be6-4b04-4629-9ff8-37362ab0a313" (UID: "81515be6-4b04-4629-9ff8-37362ab0a313"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.520861 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.520890 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct6mw\" (UniqueName: \"kubernetes.io/projected/81515be6-4b04-4629-9ff8-37362ab0a313-kube-api-access-ct6mw\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.520921 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.520930 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.543596 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "81515be6-4b04-4629-9ff8-37362ab0a313" (UID: "81515be6-4b04-4629-9ff8-37362ab0a313"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.622968 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81515be6-4b04-4629-9ff8-37362ab0a313-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.822869 4838 generic.go:334] "Generic (PLEG): container finished" podID="81515be6-4b04-4629-9ff8-37362ab0a313" containerID="3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2" exitCode=0 Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.822945 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.822973 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" event={"ID":"81515be6-4b04-4629-9ff8-37362ab0a313","Type":"ContainerDied","Data":"3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2"} Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.823381 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-7f9tn" event={"ID":"81515be6-4b04-4629-9ff8-37362ab0a313","Type":"ContainerDied","Data":"4cbc7c21fbef34268e6cb1151b33366027d44d2ef25641be81fd256f43acc38b"} Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.823403 4838 scope.go:117] "RemoveContainer" containerID="3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.829964 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerStarted","Data":"2ff1ac5ffa2469eea370484ef5e53c553c70cdb92d241f06ebc96d0c92f4b517"} Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.847429 4838 scope.go:117] "RemoveContainer" containerID="edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.875945 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-7f9tn"] Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.890344 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-7f9tn"] Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.891450 4838 scope.go:117] "RemoveContainer" containerID="3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2" Dec 07 09:28:20 crc kubenswrapper[4838]: E1207 09:28:20.891772 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2\": container with ID starting with 3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2 not found: ID does not exist" containerID="3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.891802 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2"} err="failed to get container status \"3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2\": rpc error: code = NotFound desc = could not find container \"3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2\": container with ID starting with 3194fd6f31b67816d6d56ee4bdb80a64556b3052859ef8193fc24c6b8827d2b2 not found: ID does not exist" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.891838 4838 scope.go:117] "RemoveContainer" containerID="edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7" Dec 07 09:28:20 crc kubenswrapper[4838]: E1207 09:28:20.892182 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7\": container with ID starting with edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7 not found: ID does not exist" containerID="edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7" Dec 07 09:28:20 crc kubenswrapper[4838]: I1207 09:28:20.892211 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7"} err="failed to get container status \"edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7\": rpc error: code = NotFound desc = could not find container \"edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7\": container with ID starting with edb9155d571302af8d1a511c2a65daff72d8f7c9e080350edc0ec53ecd533cd7 not found: ID does not exist" Dec 07 09:28:21 crc kubenswrapper[4838]: I1207 09:28:21.623084 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81515be6-4b04-4629-9ff8-37362ab0a313" path="/var/lib/kubelet/pods/81515be6-4b04-4629-9ff8-37362ab0a313/volumes" Dec 07 09:28:21 crc kubenswrapper[4838]: I1207 09:28:21.845243 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerStarted","Data":"749c49de29181a11711d0234e8033adb4b780ef8ea3126f59efcbd244967798e"} Dec 07 09:28:21 crc kubenswrapper[4838]: I1207 09:28:21.845440 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 09:28:24 crc kubenswrapper[4838]: I1207 09:28:24.492860 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:28:24 crc kubenswrapper[4838]: I1207 09:28:24.493465 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:28:24 crc kubenswrapper[4838]: I1207 09:28:24.876629 4838 generic.go:334] "Generic (PLEG): container finished" podID="e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" containerID="b25b83605bcbe9e2002ca0fe388109c280cc08dd827f6dc0b00c740c1151d4ca" exitCode=0 Dec 07 09:28:24 crc kubenswrapper[4838]: I1207 09:28:24.876684 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wndph" event={"ID":"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137","Type":"ContainerDied","Data":"b25b83605bcbe9e2002ca0fe388109c280cc08dd827f6dc0b00c740c1151d4ca"} Dec 07 09:28:24 crc kubenswrapper[4838]: I1207 09:28:24.896409 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.270670323 podStartE2EDuration="8.896385937s" podCreationTimestamp="2025-12-07 09:28:16 +0000 UTC" firstStartedPulling="2025-12-07 09:28:17.632733597 +0000 UTC m=+1314.340052614" lastFinishedPulling="2025-12-07 09:28:21.258449211 +0000 UTC m=+1317.965768228" observedRunningTime="2025-12-07 09:28:21.869505529 +0000 UTC m=+1318.576824546" watchObservedRunningTime="2025-12-07 09:28:24.896385937 +0000 UTC m=+1321.603704994" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.075865 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.076213 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.230688 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.329550 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-config-data\") pod \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.329630 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-combined-ca-bundle\") pod \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.329732 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2nbx\" (UniqueName: \"kubernetes.io/projected/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-kube-api-access-f2nbx\") pod \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.329806 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-scripts\") pod \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\" (UID: \"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137\") " Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.339009 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-scripts" (OuterVolumeSpecName: "scripts") pod "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" (UID: "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.356010 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-kube-api-access-f2nbx" (OuterVolumeSpecName: "kube-api-access-f2nbx") pod "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" (UID: "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137"). InnerVolumeSpecName "kube-api-access-f2nbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.365584 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-config-data" (OuterVolumeSpecName: "config-data") pod "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" (UID: "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.367883 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" (UID: "e9fd40d9-0ac7-47d6-afd7-05a6a88ef137"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.432105 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2nbx\" (UniqueName: \"kubernetes.io/projected/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-kube-api-access-f2nbx\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.432140 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.432149 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.432158 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.901139 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-wndph" event={"ID":"e9fd40d9-0ac7-47d6-afd7-05a6a88ef137","Type":"ContainerDied","Data":"b8b2878ae0f6117d792fa811d8536920cdcbca3c360be4e8674e1536982ec952"} Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.901201 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8b2878ae0f6117d792fa811d8536920cdcbca3c360be4e8674e1536982ec952" Dec 07 09:28:26 crc kubenswrapper[4838]: I1207 09:28:26.901782 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-wndph" Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.084639 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.085221 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-log" containerID="cri-o://02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5" gracePeriod=30 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.085656 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-api" containerID="cri-o://95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8" gracePeriod=30 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.093064 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.178:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.093152 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.178:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.125992 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.126240 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" containerName="nova-scheduler-scheduler" containerID="cri-o://22b43f0b83070e3c6e92d8f2c0e7f80d9f9dbe3dbcd9f39a068578d945a048ec" gracePeriod=30 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.152865 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.153099 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-log" containerID="cri-o://f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c" gracePeriod=30 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.153375 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-metadata" containerID="cri-o://089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18" gracePeriod=30 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.911341 4838 generic.go:334] "Generic (PLEG): container finished" podID="4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" containerID="22b43f0b83070e3c6e92d8f2c0e7f80d9f9dbe3dbcd9f39a068578d945a048ec" exitCode=0 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.911425 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0","Type":"ContainerDied","Data":"22b43f0b83070e3c6e92d8f2c0e7f80d9f9dbe3dbcd9f39a068578d945a048ec"} Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.913214 4838 generic.go:334] "Generic (PLEG): container finished" podID="30ca3256-1182-42cc-862f-58d99bec1b42" containerID="02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5" exitCode=143 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.913282 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30ca3256-1182-42cc-862f-58d99bec1b42","Type":"ContainerDied","Data":"02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5"} Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.914727 4838 generic.go:334] "Generic (PLEG): container finished" podID="d8a727b0-faa8-499a-bb85-c6234343f062" containerID="f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c" exitCode=143 Dec 07 09:28:27 crc kubenswrapper[4838]: I1207 09:28:27.914751 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d8a727b0-faa8-499a-bb85-c6234343f062","Type":"ContainerDied","Data":"f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c"} Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.197101 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.263863 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-config-data\") pod \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.263957 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-combined-ca-bundle\") pod \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.263993 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drhh6\" (UniqueName: \"kubernetes.io/projected/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-kube-api-access-drhh6\") pod \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\" (UID: \"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0\") " Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.282547 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-kube-api-access-drhh6" (OuterVolumeSpecName: "kube-api-access-drhh6") pod "4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" (UID: "4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0"). InnerVolumeSpecName "kube-api-access-drhh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.300399 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" (UID: "4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.328199 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-config-data" (OuterVolumeSpecName: "config-data") pod "4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" (UID: "4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.365673 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.365952 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drhh6\" (UniqueName: \"kubernetes.io/projected/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-kube-api-access-drhh6\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.366031 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.924731 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0","Type":"ContainerDied","Data":"c307aafc0971e53d85ced2f15b30bda49ea0457ae40428ede38eba57e6320cb1"} Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.924838 4838 scope.go:117] "RemoveContainer" containerID="22b43f0b83070e3c6e92d8f2c0e7f80d9f9dbe3dbcd9f39a068578d945a048ec" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.924840 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.962779 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.973296 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.991951 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:28:28 crc kubenswrapper[4838]: E1207 09:28:28.992483 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81515be6-4b04-4629-9ff8-37362ab0a313" containerName="dnsmasq-dns" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.992507 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="81515be6-4b04-4629-9ff8-37362ab0a313" containerName="dnsmasq-dns" Dec 07 09:28:28 crc kubenswrapper[4838]: E1207 09:28:28.992533 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81515be6-4b04-4629-9ff8-37362ab0a313" containerName="init" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.992541 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="81515be6-4b04-4629-9ff8-37362ab0a313" containerName="init" Dec 07 09:28:28 crc kubenswrapper[4838]: E1207 09:28:28.992568 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" containerName="nova-manage" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.992577 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" containerName="nova-manage" Dec 07 09:28:28 crc kubenswrapper[4838]: E1207 09:28:28.992587 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" containerName="nova-scheduler-scheduler" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.992595 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" containerName="nova-scheduler-scheduler" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.992795 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" containerName="nova-scheduler-scheduler" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.992828 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="81515be6-4b04-4629-9ff8-37362ab0a313" containerName="dnsmasq-dns" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.992847 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" containerName="nova-manage" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.993609 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.996771 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 07 09:28:28 crc kubenswrapper[4838]: I1207 09:28:28.999088 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.082562 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f3b4db3-4d04-4198-972d-ab3722b30abe-config-data\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.082788 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpm2r\" (UniqueName: \"kubernetes.io/projected/0f3b4db3-4d04-4198-972d-ab3722b30abe-kube-api-access-rpm2r\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.083019 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f3b4db3-4d04-4198-972d-ab3722b30abe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.184174 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpm2r\" (UniqueName: \"kubernetes.io/projected/0f3b4db3-4d04-4198-972d-ab3722b30abe-kube-api-access-rpm2r\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.184276 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f3b4db3-4d04-4198-972d-ab3722b30abe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.184313 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f3b4db3-4d04-4198-972d-ab3722b30abe-config-data\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.190196 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f3b4db3-4d04-4198-972d-ab3722b30abe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.206912 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f3b4db3-4d04-4198-972d-ab3722b30abe-config-data\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.213925 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpm2r\" (UniqueName: \"kubernetes.io/projected/0f3b4db3-4d04-4198-972d-ab3722b30abe-kube-api-access-rpm2r\") pod \"nova-scheduler-0\" (UID: \"0f3b4db3-4d04-4198-972d-ab3722b30abe\") " pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.323630 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.624692 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0" path="/var/lib/kubelet/pods/4d27a2f0-b17d-4d7b-9342-e5b5e22b8da0/volumes" Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.826267 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 09:28:29 crc kubenswrapper[4838]: I1207 09:28:29.934108 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f3b4db3-4d04-4198-972d-ab3722b30abe","Type":"ContainerStarted","Data":"0bc7ae2b3db54a43ba5f9f28e17aca8a7297f4a1d1fb794f3711b5330fd97d0d"} Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.687042 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.717965 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-combined-ca-bundle\") pod \"d8a727b0-faa8-499a-bb85-c6234343f062\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.718029 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-nova-metadata-tls-certs\") pod \"d8a727b0-faa8-499a-bb85-c6234343f062\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.718060 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a727b0-faa8-499a-bb85-c6234343f062-logs\") pod \"d8a727b0-faa8-499a-bb85-c6234343f062\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.718192 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-config-data\") pod \"d8a727b0-faa8-499a-bb85-c6234343f062\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.718252 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62drd\" (UniqueName: \"kubernetes.io/projected/d8a727b0-faa8-499a-bb85-c6234343f062-kube-api-access-62drd\") pod \"d8a727b0-faa8-499a-bb85-c6234343f062\" (UID: \"d8a727b0-faa8-499a-bb85-c6234343f062\") " Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.720176 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8a727b0-faa8-499a-bb85-c6234343f062-logs" (OuterVolumeSpecName: "logs") pod "d8a727b0-faa8-499a-bb85-c6234343f062" (UID: "d8a727b0-faa8-499a-bb85-c6234343f062"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.725078 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8a727b0-faa8-499a-bb85-c6234343f062-kube-api-access-62drd" (OuterVolumeSpecName: "kube-api-access-62drd") pod "d8a727b0-faa8-499a-bb85-c6234343f062" (UID: "d8a727b0-faa8-499a-bb85-c6234343f062"). InnerVolumeSpecName "kube-api-access-62drd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.760866 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-config-data" (OuterVolumeSpecName: "config-data") pod "d8a727b0-faa8-499a-bb85-c6234343f062" (UID: "d8a727b0-faa8-499a-bb85-c6234343f062"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.797961 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d8a727b0-faa8-499a-bb85-c6234343f062" (UID: "d8a727b0-faa8-499a-bb85-c6234343f062"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.798212 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8a727b0-faa8-499a-bb85-c6234343f062" (UID: "d8a727b0-faa8-499a-bb85-c6234343f062"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.820688 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.820716 4838 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.820726 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8a727b0-faa8-499a-bb85-c6234343f062-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.820734 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a727b0-faa8-499a-bb85-c6234343f062-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.820742 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62drd\" (UniqueName: \"kubernetes.io/projected/d8a727b0-faa8-499a-bb85-c6234343f062-kube-api-access-62drd\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.949193 4838 generic.go:334] "Generic (PLEG): container finished" podID="d8a727b0-faa8-499a-bb85-c6234343f062" containerID="089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18" exitCode=0 Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.949492 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d8a727b0-faa8-499a-bb85-c6234343f062","Type":"ContainerDied","Data":"089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18"} Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.949627 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d8a727b0-faa8-499a-bb85-c6234343f062","Type":"ContainerDied","Data":"defebda0ffdcc66f198c1af6507bacf5e033e5bfa6a0f9d03dbc78068a5ab98f"} Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.949720 4838 scope.go:117] "RemoveContainer" containerID="089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.949947 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.958045 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f3b4db3-4d04-4198-972d-ab3722b30abe","Type":"ContainerStarted","Data":"a9fddb0ff80e36701530070561cf82152ef51739905f1b9953fe35ee2efc938e"} Dec 07 09:28:30 crc kubenswrapper[4838]: I1207 09:28:30.990801 4838 scope.go:117] "RemoveContainer" containerID="f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.000769 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.000720114 podStartE2EDuration="3.000720114s" podCreationTimestamp="2025-12-07 09:28:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:28:30.985000891 +0000 UTC m=+1327.692319908" watchObservedRunningTime="2025-12-07 09:28:31.000720114 +0000 UTC m=+1327.708039171" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.026666 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.037849 4838 scope.go:117] "RemoveContainer" containerID="089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.037912 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:28:31 crc kubenswrapper[4838]: E1207 09:28:31.038306 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18\": container with ID starting with 089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18 not found: ID does not exist" containerID="089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.038340 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18"} err="failed to get container status \"089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18\": rpc error: code = NotFound desc = could not find container \"089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18\": container with ID starting with 089923a27cf111d489a6b907a17992e1fd31f663566ef25a3c34011e0fcc1d18 not found: ID does not exist" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.038364 4838 scope.go:117] "RemoveContainer" containerID="f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c" Dec 07 09:28:31 crc kubenswrapper[4838]: E1207 09:28:31.038647 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c\": container with ID starting with f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c not found: ID does not exist" containerID="f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.038680 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c"} err="failed to get container status \"f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c\": rpc error: code = NotFound desc = could not find container \"f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c\": container with ID starting with f43f67518ff2e71a49ca027bb69ff64a3ed34c3bac9febf5b8c1ffb22f18411c not found: ID does not exist" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.047024 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:28:31 crc kubenswrapper[4838]: E1207 09:28:31.047486 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-metadata" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.047509 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-metadata" Dec 07 09:28:31 crc kubenswrapper[4838]: E1207 09:28:31.047521 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-log" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.047528 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-log" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.047798 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-metadata" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.047844 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" containerName="nova-metadata-log" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.048941 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.057573 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.059559 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.069908 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.129608 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.129678 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.129708 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg9xt\" (UniqueName: \"kubernetes.io/projected/021e33aa-1082-46ef-afa0-1dd04aad6502-kube-api-access-mg9xt\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.129728 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-config-data\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.129763 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/021e33aa-1082-46ef-afa0-1dd04aad6502-logs\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.231855 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.232285 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.232318 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg9xt\" (UniqueName: \"kubernetes.io/projected/021e33aa-1082-46ef-afa0-1dd04aad6502-kube-api-access-mg9xt\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.232343 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-config-data\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.232399 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/021e33aa-1082-46ef-afa0-1dd04aad6502-logs\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.233029 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/021e33aa-1082-46ef-afa0-1dd04aad6502-logs\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.238133 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-config-data\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.238793 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.240413 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/021e33aa-1082-46ef-afa0-1dd04aad6502-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.256727 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg9xt\" (UniqueName: \"kubernetes.io/projected/021e33aa-1082-46ef-afa0-1dd04aad6502-kube-api-access-mg9xt\") pod \"nova-metadata-0\" (UID: \"021e33aa-1082-46ef-afa0-1dd04aad6502\") " pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.387417 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.629313 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8a727b0-faa8-499a-bb85-c6234343f062" path="/var/lib/kubelet/pods/d8a727b0-faa8-499a-bb85-c6234343f062/volumes" Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.874370 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 09:28:31 crc kubenswrapper[4838]: I1207 09:28:31.969097 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"021e33aa-1082-46ef-afa0-1dd04aad6502","Type":"ContainerStarted","Data":"bc8b6ae23ef5ecea50074d7c1ad1f0f07425d17c99cd34a91e90582f4686ad10"} Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.965176 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.982085 4838 generic.go:334] "Generic (PLEG): container finished" podID="30ca3256-1182-42cc-862f-58d99bec1b42" containerID="95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8" exitCode=0 Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.982149 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.982150 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30ca3256-1182-42cc-862f-58d99bec1b42","Type":"ContainerDied","Data":"95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8"} Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.982382 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30ca3256-1182-42cc-862f-58d99bec1b42","Type":"ContainerDied","Data":"ba464b85a3a20d697683de69b8fded68b4491d17b30c75a98897136ec46a1ced"} Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.982412 4838 scope.go:117] "RemoveContainer" containerID="95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8" Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.985359 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"021e33aa-1082-46ef-afa0-1dd04aad6502","Type":"ContainerStarted","Data":"691242b867a4fc668cde5ca7f667ef7b626e6b3d42d14fc9ad5f00ba9cea208c"} Dec 07 09:28:32 crc kubenswrapper[4838]: I1207 09:28:32.985405 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"021e33aa-1082-46ef-afa0-1dd04aad6502","Type":"ContainerStarted","Data":"ef9503b2fd90b76efcaa583db220f45b6d7e46ce693b26c0f23cd0f0094bdfee"} Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.016808 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.016785401 podStartE2EDuration="2.016785401s" podCreationTimestamp="2025-12-07 09:28:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:28:33.010205138 +0000 UTC m=+1329.717524165" watchObservedRunningTime="2025-12-07 09:28:33.016785401 +0000 UTC m=+1329.724104418" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.024236 4838 scope.go:117] "RemoveContainer" containerID="02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.040576 4838 scope.go:117] "RemoveContainer" containerID="95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8" Dec 07 09:28:33 crc kubenswrapper[4838]: E1207 09:28:33.041005 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8\": container with ID starting with 95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8 not found: ID does not exist" containerID="95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.041046 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8"} err="failed to get container status \"95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8\": rpc error: code = NotFound desc = could not find container \"95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8\": container with ID starting with 95fc1a5cadbd2972d30e43ede77da5b8a9b01938d0e0d7e7bde64b1f4fb0feb8 not found: ID does not exist" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.041071 4838 scope.go:117] "RemoveContainer" containerID="02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5" Dec 07 09:28:33 crc kubenswrapper[4838]: E1207 09:28:33.041319 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5\": container with ID starting with 02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5 not found: ID does not exist" containerID="02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.041356 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5"} err="failed to get container status \"02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5\": rpc error: code = NotFound desc = could not find container \"02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5\": container with ID starting with 02beae16c79bf637702a33d3917df7bab95ef2673c775b74bf368ae85c1aa6a5 not found: ID does not exist" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.068260 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ca3256-1182-42cc-862f-58d99bec1b42-logs\") pod \"30ca3256-1182-42cc-862f-58d99bec1b42\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.068324 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-config-data\") pod \"30ca3256-1182-42cc-862f-58d99bec1b42\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.068363 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-combined-ca-bundle\") pod \"30ca3256-1182-42cc-862f-58d99bec1b42\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.068414 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-public-tls-certs\") pod \"30ca3256-1182-42cc-862f-58d99bec1b42\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.068482 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlvvv\" (UniqueName: \"kubernetes.io/projected/30ca3256-1182-42cc-862f-58d99bec1b42-kube-api-access-zlvvv\") pod \"30ca3256-1182-42cc-862f-58d99bec1b42\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.068577 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-internal-tls-certs\") pod \"30ca3256-1182-42cc-862f-58d99bec1b42\" (UID: \"30ca3256-1182-42cc-862f-58d99bec1b42\") " Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.069088 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30ca3256-1182-42cc-862f-58d99bec1b42-logs" (OuterVolumeSpecName: "logs") pod "30ca3256-1182-42cc-862f-58d99bec1b42" (UID: "30ca3256-1182-42cc-862f-58d99bec1b42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.074875 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30ca3256-1182-42cc-862f-58d99bec1b42-kube-api-access-zlvvv" (OuterVolumeSpecName: "kube-api-access-zlvvv") pod "30ca3256-1182-42cc-862f-58d99bec1b42" (UID: "30ca3256-1182-42cc-862f-58d99bec1b42"). InnerVolumeSpecName "kube-api-access-zlvvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.098370 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-config-data" (OuterVolumeSpecName: "config-data") pod "30ca3256-1182-42cc-862f-58d99bec1b42" (UID: "30ca3256-1182-42cc-862f-58d99bec1b42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.105304 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30ca3256-1182-42cc-862f-58d99bec1b42" (UID: "30ca3256-1182-42cc-862f-58d99bec1b42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.120181 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "30ca3256-1182-42cc-862f-58d99bec1b42" (UID: "30ca3256-1182-42cc-862f-58d99bec1b42"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.126166 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "30ca3256-1182-42cc-862f-58d99bec1b42" (UID: "30ca3256-1182-42cc-862f-58d99bec1b42"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.171187 4838 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.171227 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ca3256-1182-42cc-862f-58d99bec1b42-logs\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.171239 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.171250 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.171265 4838 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ca3256-1182-42cc-862f-58d99bec1b42-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.171276 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlvvv\" (UniqueName: \"kubernetes.io/projected/30ca3256-1182-42cc-862f-58d99bec1b42-kube-api-access-zlvvv\") on node \"crc\" DevicePath \"\"" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.330867 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.353005 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.377794 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:33 crc kubenswrapper[4838]: E1207 09:28:33.378449 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-log" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.378467 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-log" Dec 07 09:28:33 crc kubenswrapper[4838]: E1207 09:28:33.378493 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-api" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.378499 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-api" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.380371 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-log" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.380408 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" containerName="nova-api-api" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.382260 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.386335 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.387077 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.387860 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.407162 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.478244 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.478324 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.478373 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s8b6\" (UniqueName: \"kubernetes.io/projected/1d429890-5092-448b-b386-36493df22443-kube-api-access-9s8b6\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.478419 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-config-data\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.478467 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d429890-5092-448b-b386-36493df22443-logs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.478504 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.580417 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s8b6\" (UniqueName: \"kubernetes.io/projected/1d429890-5092-448b-b386-36493df22443-kube-api-access-9s8b6\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.580846 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-config-data\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.580913 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d429890-5092-448b-b386-36493df22443-logs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.581505 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d429890-5092-448b-b386-36493df22443-logs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.581599 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.582165 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.582552 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.586241 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.589494 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.589977 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.593055 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d429890-5092-448b-b386-36493df22443-config-data\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.595150 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s8b6\" (UniqueName: \"kubernetes.io/projected/1d429890-5092-448b-b386-36493df22443-kube-api-access-9s8b6\") pod \"nova-api-0\" (UID: \"1d429890-5092-448b-b386-36493df22443\") " pod="openstack/nova-api-0" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.628606 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30ca3256-1182-42cc-862f-58d99bec1b42" path="/var/lib/kubelet/pods/30ca3256-1182-42cc-862f-58d99bec1b42/volumes" Dec 07 09:28:33 crc kubenswrapper[4838]: I1207 09:28:33.748637 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 09:28:34 crc kubenswrapper[4838]: I1207 09:28:34.230447 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 09:28:34 crc kubenswrapper[4838]: W1207 09:28:34.236689 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d429890_5092_448b_b386_36493df22443.slice/crio-971d89b556688305da8d623c8ad7b121ecbe3e0dbdd048acaf62a0e6726f9826 WatchSource:0}: Error finding container 971d89b556688305da8d623c8ad7b121ecbe3e0dbdd048acaf62a0e6726f9826: Status 404 returned error can't find the container with id 971d89b556688305da8d623c8ad7b121ecbe3e0dbdd048acaf62a0e6726f9826 Dec 07 09:28:34 crc kubenswrapper[4838]: I1207 09:28:34.324061 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 07 09:28:35 crc kubenswrapper[4838]: I1207 09:28:35.010240 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d429890-5092-448b-b386-36493df22443","Type":"ContainerStarted","Data":"d7dbefdb771fa0a5397d8569f42b67214b55b85a3fa35b8136c4c5f9e0251957"} Dec 07 09:28:35 crc kubenswrapper[4838]: I1207 09:28:35.010294 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d429890-5092-448b-b386-36493df22443","Type":"ContainerStarted","Data":"a5bdfe52a475478e178c037cafc7f8c9b39d6d6c3e56fa88806a7246542d7ee1"} Dec 07 09:28:35 crc kubenswrapper[4838]: I1207 09:28:35.010315 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d429890-5092-448b-b386-36493df22443","Type":"ContainerStarted","Data":"971d89b556688305da8d623c8ad7b121ecbe3e0dbdd048acaf62a0e6726f9826"} Dec 07 09:28:36 crc kubenswrapper[4838]: I1207 09:28:36.388041 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 09:28:36 crc kubenswrapper[4838]: I1207 09:28:36.388451 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 09:28:39 crc kubenswrapper[4838]: I1207 09:28:39.324334 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 07 09:28:39 crc kubenswrapper[4838]: I1207 09:28:39.367724 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 07 09:28:39 crc kubenswrapper[4838]: I1207 09:28:39.393110 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=6.393089845 podStartE2EDuration="6.393089845s" podCreationTimestamp="2025-12-07 09:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:28:35.047729819 +0000 UTC m=+1331.755048846" watchObservedRunningTime="2025-12-07 09:28:39.393089845 +0000 UTC m=+1336.100408872" Dec 07 09:28:40 crc kubenswrapper[4838]: I1207 09:28:40.098196 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 07 09:28:41 crc kubenswrapper[4838]: I1207 09:28:41.388383 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 09:28:41 crc kubenswrapper[4838]: I1207 09:28:41.388944 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 09:28:42 crc kubenswrapper[4838]: I1207 09:28:42.405078 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="021e33aa-1082-46ef-afa0-1dd04aad6502" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:42 crc kubenswrapper[4838]: I1207 09:28:42.405093 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="021e33aa-1082-46ef-afa0-1dd04aad6502" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:43 crc kubenswrapper[4838]: I1207 09:28:43.749097 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:28:43 crc kubenswrapper[4838]: I1207 09:28:43.749435 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 09:28:44 crc kubenswrapper[4838]: I1207 09:28:44.763977 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1d429890-5092-448b-b386-36493df22443" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:44 crc kubenswrapper[4838]: I1207 09:28:44.764238 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1d429890-5092-448b-b386-36493df22443" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.183:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 09:28:47 crc kubenswrapper[4838]: I1207 09:28:47.174060 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 07 09:28:51 crc kubenswrapper[4838]: I1207 09:28:51.415438 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 09:28:51 crc kubenswrapper[4838]: I1207 09:28:51.420762 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 09:28:51 crc kubenswrapper[4838]: I1207 09:28:51.422649 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 09:28:52 crc kubenswrapper[4838]: I1207 09:28:52.170926 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 09:28:53 crc kubenswrapper[4838]: I1207 09:28:53.797255 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 09:28:53 crc kubenswrapper[4838]: I1207 09:28:53.799223 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 09:28:53 crc kubenswrapper[4838]: I1207 09:28:53.802080 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 09:28:53 crc kubenswrapper[4838]: I1207 09:28:53.832424 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 09:28:54 crc kubenswrapper[4838]: I1207 09:28:54.186962 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 09:28:54 crc kubenswrapper[4838]: I1207 09:28:54.202110 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 09:28:54 crc kubenswrapper[4838]: I1207 09:28:54.493287 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:28:54 crc kubenswrapper[4838]: I1207 09:28:54.493357 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:28:54 crc kubenswrapper[4838]: I1207 09:28:54.493414 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:28:54 crc kubenswrapper[4838]: I1207 09:28:54.494267 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:28:54 crc kubenswrapper[4838]: I1207 09:28:54.494334 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9" gracePeriod=600 Dec 07 09:28:54 crc kubenswrapper[4838]: E1207 09:28:54.746273 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1304517_3ac2_4aa3_afe2_01e27f87d880.slice/crio-1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1304517_3ac2_4aa3_afe2_01e27f87d880.slice/crio-conmon-1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9.scope\": RecentStats: unable to find data in memory cache]" Dec 07 09:28:55 crc kubenswrapper[4838]: I1207 09:28:55.197452 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9" exitCode=0 Dec 07 09:28:55 crc kubenswrapper[4838]: I1207 09:28:55.197609 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9"} Dec 07 09:28:55 crc kubenswrapper[4838]: I1207 09:28:55.198100 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3"} Dec 07 09:28:55 crc kubenswrapper[4838]: I1207 09:28:55.198154 4838 scope.go:117] "RemoveContainer" containerID="ed0d936715e5d0425ff8433d1f07f45e763ba568568132ddb5adcdb73cc5a936" Dec 07 09:29:01 crc kubenswrapper[4838]: I1207 09:29:01.931749 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:29:03 crc kubenswrapper[4838]: I1207 09:29:03.635056 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:29:06 crc kubenswrapper[4838]: I1207 09:29:06.556093 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerName="rabbitmq" containerID="cri-o://269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004" gracePeriod=604796 Dec 07 09:29:08 crc kubenswrapper[4838]: I1207 09:29:08.500141 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerName="rabbitmq" containerID="cri-o://a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03" gracePeriod=604796 Dec 07 09:29:09 crc kubenswrapper[4838]: I1207 09:29:09.427219 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 07 09:29:09 crc kubenswrapper[4838]: I1207 09:29:09.883426 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.157562 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225075 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-plugins\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225123 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-confd\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225152 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-plugins-conf\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225208 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-config-data\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225246 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eae80b29-4c64-403b-be38-7fd30b7ba66f-erlang-cookie-secret\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225292 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225346 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-tls\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225526 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eae80b29-4c64-403b-be38-7fd30b7ba66f-pod-info\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225564 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4pn6\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-kube-api-access-m4pn6\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225594 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-server-conf\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225628 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-erlang-cookie\") pod \"eae80b29-4c64-403b-be38-7fd30b7ba66f\" (UID: \"eae80b29-4c64-403b-be38-7fd30b7ba66f\") " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.225767 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.226312 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.226876 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.226911 4838 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.256791 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.265006 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.269450 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae80b29-4c64-403b-be38-7fd30b7ba66f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.271102 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-kube-api-access-m4pn6" (OuterVolumeSpecName: "kube-api-access-m4pn6") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "kube-api-access-m4pn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.272114 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.273112 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/eae80b29-4c64-403b-be38-7fd30b7ba66f-pod-info" (OuterVolumeSpecName: "pod-info") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.330100 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.330471 4838 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eae80b29-4c64-403b-be38-7fd30b7ba66f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.330591 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4pn6\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-kube-api-access-m4pn6\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.330676 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.330869 4838 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eae80b29-4c64-403b-be38-7fd30b7ba66f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.330981 4838 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.360042 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-server-conf" (OuterVolumeSpecName: "server-conf") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.378538 4838 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.384145 4838 generic.go:334] "Generic (PLEG): container finished" podID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerID="269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004" exitCode=0 Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.384192 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eae80b29-4c64-403b-be38-7fd30b7ba66f","Type":"ContainerDied","Data":"269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004"} Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.384223 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eae80b29-4c64-403b-be38-7fd30b7ba66f","Type":"ContainerDied","Data":"19de77fd1c4624bfd4eab30715225b42a8401b4eb1bbdeaf0aa3b4cc814e2316"} Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.384242 4838 scope.go:117] "RemoveContainer" containerID="269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.384434 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.393422 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-config-data" (OuterVolumeSpecName: "config-data") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.425588 4838 scope.go:117] "RemoveContainer" containerID="73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.432579 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.432611 4838 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.432619 4838 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eae80b29-4c64-403b-be38-7fd30b7ba66f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.460837 4838 scope.go:117] "RemoveContainer" containerID="269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004" Dec 07 09:29:13 crc kubenswrapper[4838]: E1207 09:29:13.461256 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004\": container with ID starting with 269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004 not found: ID does not exist" containerID="269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.461284 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004"} err="failed to get container status \"269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004\": rpc error: code = NotFound desc = could not find container \"269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004\": container with ID starting with 269853f314ff6b8f5fe4967f9059dc61ab78a34ad00fbf093950044bddef0004 not found: ID does not exist" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.461303 4838 scope.go:117] "RemoveContainer" containerID="73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2" Dec 07 09:29:13 crc kubenswrapper[4838]: E1207 09:29:13.461677 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2\": container with ID starting with 73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2 not found: ID does not exist" containerID="73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.461699 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2"} err="failed to get container status \"73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2\": rpc error: code = NotFound desc = could not find container \"73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2\": container with ID starting with 73a3594fa99e5e79dc41a2eb904f0b0663634bb49cdc94ee3a0c3d40bf04abb2 not found: ID does not exist" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.481483 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "eae80b29-4c64-403b-be38-7fd30b7ba66f" (UID: "eae80b29-4c64-403b-be38-7fd30b7ba66f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.534335 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eae80b29-4c64-403b-be38-7fd30b7ba66f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.710004 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.717392 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.738632 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:29:13 crc kubenswrapper[4838]: E1207 09:29:13.738981 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerName="setup-container" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.738997 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerName="setup-container" Dec 07 09:29:13 crc kubenswrapper[4838]: E1207 09:29:13.739026 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerName="rabbitmq" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.739032 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerName="rabbitmq" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.739201 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" containerName="rabbitmq" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.740075 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.743354 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.743397 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.743534 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nxdd4" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.744000 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.746376 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.746427 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.746569 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.822788 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839105 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839171 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl5gt\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-kube-api-access-vl5gt\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839238 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839285 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839322 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839352 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839487 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/688c3086-fd09-4b9d-921e-e41c7f8a01ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839523 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839633 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839677 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/688c3086-fd09-4b9d-921e-e41c7f8a01ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.839744 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.941953 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942031 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/688c3086-fd09-4b9d-921e-e41c7f8a01ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942067 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942094 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942116 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl5gt\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-kube-api-access-vl5gt\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942143 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942176 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942211 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942239 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942277 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/688c3086-fd09-4b9d-921e-e41c7f8a01ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.942296 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.943736 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.944415 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.945186 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.945331 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.945356 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.946716 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/688c3086-fd09-4b9d-921e-e41c7f8a01ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.950463 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/688c3086-fd09-4b9d-921e-e41c7f8a01ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.950957 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/688c3086-fd09-4b9d-921e-e41c7f8a01ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.951323 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.954654 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.966111 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl5gt\" (UniqueName: \"kubernetes.io/projected/688c3086-fd09-4b9d-921e-e41c7f8a01ac-kube-api-access-vl5gt\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:13 crc kubenswrapper[4838]: I1207 09:29:13.992613 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"688c3086-fd09-4b9d-921e-e41c7f8a01ac\") " pod="openstack/rabbitmq-server-0" Dec 07 09:29:14 crc kubenswrapper[4838]: I1207 09:29:14.061926 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 09:29:14 crc kubenswrapper[4838]: I1207 09:29:14.483040 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.035778 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074296 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-erlang-cookie\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074364 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-plugins\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074403 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-confd\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074430 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074453 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-pod-info\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074488 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-plugins-conf\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074525 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-config-data\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074562 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-tls\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074588 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzmbr\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-kube-api-access-qzmbr\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074638 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-erlang-cookie-secret\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.074681 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-server-conf\") pod \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\" (UID: \"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454\") " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.079240 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.079591 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.094398 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.095007 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.101384 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-pod-info" (OuterVolumeSpecName: "pod-info") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.108935 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.139033 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-kube-api-access-qzmbr" (OuterVolumeSpecName: "kube-api-access-qzmbr") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "kube-api-access-qzmbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.141727 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.176960 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.177006 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.177040 4838 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.177054 4838 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-pod-info\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.177065 4838 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.177078 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.177089 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzmbr\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-kube-api-access-qzmbr\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.177101 4838 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.221741 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-config-data" (OuterVolumeSpecName: "config-data") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.256733 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-server-conf" (OuterVolumeSpecName: "server-conf") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.271650 4838 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.279983 4838 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.280019 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.280032 4838 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-server-conf\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.330181 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" (UID: "20d0a08e-11bd-4cb9-b5c7-b61abbbd4454"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.381999 4838 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.436139 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"688c3086-fd09-4b9d-921e-e41c7f8a01ac","Type":"ContainerStarted","Data":"9dde684a9708c6ebc73d7a489db38b45a5cd2145a078752db5123f2bbd8ad7dd"} Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.438433 4838 generic.go:334] "Generic (PLEG): container finished" podID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerID="a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03" exitCode=0 Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.438619 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.438611 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454","Type":"ContainerDied","Data":"a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03"} Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.438930 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"20d0a08e-11bd-4cb9-b5c7-b61abbbd4454","Type":"ContainerDied","Data":"1684f067ea189296b647a867168cc7d63950375ad7c8de7aaeac4327c4a6d034"} Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.439036 4838 scope.go:117] "RemoveContainer" containerID="a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.460711 4838 scope.go:117] "RemoveContainer" containerID="fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.474259 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.482565 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.502598 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:29:15 crc kubenswrapper[4838]: E1207 09:29:15.503844 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerName="rabbitmq" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.503929 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerName="rabbitmq" Dec 07 09:29:15 crc kubenswrapper[4838]: E1207 09:29:15.503985 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerName="setup-container" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.504033 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerName="setup-container" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.504252 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" containerName="rabbitmq" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.505147 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.511641 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.511693 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.511651 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.511909 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.512037 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.513446 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-875vw" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.517672 4838 scope.go:117] "RemoveContainer" containerID="a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03" Dec 07 09:29:15 crc kubenswrapper[4838]: E1207 09:29:15.521530 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03\": container with ID starting with a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03 not found: ID does not exist" containerID="a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.521579 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03"} err="failed to get container status \"a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03\": rpc error: code = NotFound desc = could not find container \"a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03\": container with ID starting with a021e4de08cef511610d1911897f1a8cd19206a4a3223b13d095035562e43a03 not found: ID does not exist" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.521605 4838 scope.go:117] "RemoveContainer" containerID="fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7" Dec 07 09:29:15 crc kubenswrapper[4838]: E1207 09:29:15.522506 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7\": container with ID starting with fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7 not found: ID does not exist" containerID="fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.522548 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7"} err="failed to get container status \"fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7\": rpc error: code = NotFound desc = could not find container \"fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7\": container with ID starting with fa90bb04a1907a2a8383af6b03f815ca1a0dfce2a6923b5982a4bbae23f187d7 not found: ID does not exist" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.524917 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.549680 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.624870 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20d0a08e-11bd-4cb9-b5c7-b61abbbd4454" path="/var/lib/kubelet/pods/20d0a08e-11bd-4cb9-b5c7-b61abbbd4454/volumes" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.625745 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae80b29-4c64-403b-be38-7fd30b7ba66f" path="/var/lib/kubelet/pods/eae80b29-4c64-403b-be38-7fd30b7ba66f/volumes" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.686579 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a017a788-cb2e-45ed-bceb-26ea6b69a02d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.686627 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.686677 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8lkj\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-kube-api-access-x8lkj\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.686826 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.686864 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.686896 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.686975 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a017a788-cb2e-45ed-bceb-26ea6b69a02d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.687017 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.687061 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.687136 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.687184 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.789220 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8lkj\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-kube-api-access-x8lkj\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790328 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790366 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790392 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790457 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a017a788-cb2e-45ed-bceb-26ea6b69a02d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790780 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790785 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790487 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.791137 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.790837 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.791171 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.791246 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.791422 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a017a788-cb2e-45ed-bceb-26ea6b69a02d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.791469 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.791488 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.792195 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.792315 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a017a788-cb2e-45ed-bceb-26ea6b69a02d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.808859 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a017a788-cb2e-45ed-bceb-26ea6b69a02d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.809204 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.809785 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a017a788-cb2e-45ed-bceb-26ea6b69a02d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.811712 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.813671 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8lkj\" (UniqueName: \"kubernetes.io/projected/a017a788-cb2e-45ed-bceb-26ea6b69a02d-kube-api-access-x8lkj\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:15 crc kubenswrapper[4838]: I1207 09:29:15.839917 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a017a788-cb2e-45ed-bceb-26ea6b69a02d\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.123651 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.393213 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9s96k"] Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.396126 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.408696 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9s96k"] Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.450013 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"688c3086-fd09-4b9d-921e-e41c7f8a01ac","Type":"ContainerStarted","Data":"8f03bef63131f4e707fc2572584aace63debb4cd682deeabffb08f14d1f33db7"} Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.503356 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-catalog-content\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.503446 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-utilities\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.504998 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gr9s\" (UniqueName: \"kubernetes.io/projected/05544f2f-de2e-4a39-9b19-79542a4d3508-kube-api-access-6gr9s\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.591001 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.606792 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gr9s\" (UniqueName: \"kubernetes.io/projected/05544f2f-de2e-4a39-9b19-79542a4d3508-kube-api-access-6gr9s\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.607040 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-catalog-content\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.607263 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-utilities\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.608755 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-utilities\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.609064 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-catalog-content\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.631162 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gr9s\" (UniqueName: \"kubernetes.io/projected/05544f2f-de2e-4a39-9b19-79542a4d3508-kube-api-access-6gr9s\") pod \"redhat-operators-9s96k\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:16 crc kubenswrapper[4838]: I1207 09:29:16.731464 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:17 crc kubenswrapper[4838]: W1207 09:29:17.186335 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05544f2f_de2e_4a39_9b19_79542a4d3508.slice/crio-2d1d6189ee79152e14ddc8787b793251d24f0611f79cd08a69cd748a67dd8255 WatchSource:0}: Error finding container 2d1d6189ee79152e14ddc8787b793251d24f0611f79cd08a69cd748a67dd8255: Status 404 returned error can't find the container with id 2d1d6189ee79152e14ddc8787b793251d24f0611f79cd08a69cd748a67dd8255 Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.188142 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9s96k"] Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.465887 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a017a788-cb2e-45ed-bceb-26ea6b69a02d","Type":"ContainerStarted","Data":"5cb17d1ebe386e82c3a7d505a17d28541a7aa2f5d7b5d55e86440585d5445e74"} Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.471433 4838 generic.go:334] "Generic (PLEG): container finished" podID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerID="e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5" exitCode=0 Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.471520 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s96k" event={"ID":"05544f2f-de2e-4a39-9b19-79542a4d3508","Type":"ContainerDied","Data":"e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5"} Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.471566 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s96k" event={"ID":"05544f2f-de2e-4a39-9b19-79542a4d3508","Type":"ContainerStarted","Data":"2d1d6189ee79152e14ddc8787b793251d24f0611f79cd08a69cd748a67dd8255"} Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.604764 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8676544467-lv5g7"] Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.607708 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.619776 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.635437 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8676544467-lv5g7"] Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.728146 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-sb\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.728236 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-config\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.728277 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4vtz\" (UniqueName: \"kubernetes.io/projected/2eafdf67-be05-41b8-b67a-7ba69fd9774f-kube-api-access-b4vtz\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.728333 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-nb\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.728361 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-dns-svc\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.728376 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-openstack-edpm-ipam\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.830126 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-sb\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.830243 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-config\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.830287 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4vtz\" (UniqueName: \"kubernetes.io/projected/2eafdf67-be05-41b8-b67a-7ba69fd9774f-kube-api-access-b4vtz\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.830346 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-nb\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.830379 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-dns-svc\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.830396 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-openstack-edpm-ipam\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.831696 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-nb\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.831716 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-sb\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.831733 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-config\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.831930 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-dns-svc\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.832201 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-openstack-edpm-ipam\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.853889 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4vtz\" (UniqueName: \"kubernetes.io/projected/2eafdf67-be05-41b8-b67a-7ba69fd9774f-kube-api-access-b4vtz\") pod \"dnsmasq-dns-8676544467-lv5g7\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:17 crc kubenswrapper[4838]: I1207 09:29:17.931986 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:18 crc kubenswrapper[4838]: I1207 09:29:18.480567 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a017a788-cb2e-45ed-bceb-26ea6b69a02d","Type":"ContainerStarted","Data":"65bddec28c029221a0a8a86d4939de10e2a37675e5f15f2e6a80fb125b3ebce6"} Dec 07 09:29:18 crc kubenswrapper[4838]: I1207 09:29:18.486466 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s96k" event={"ID":"05544f2f-de2e-4a39-9b19-79542a4d3508","Type":"ContainerStarted","Data":"080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128"} Dec 07 09:29:18 crc kubenswrapper[4838]: I1207 09:29:18.621494 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8676544467-lv5g7"] Dec 07 09:29:19 crc kubenswrapper[4838]: I1207 09:29:19.496234 4838 generic.go:334] "Generic (PLEG): container finished" podID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerID="b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1" exitCode=0 Dec 07 09:29:19 crc kubenswrapper[4838]: I1207 09:29:19.498134 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-lv5g7" event={"ID":"2eafdf67-be05-41b8-b67a-7ba69fd9774f","Type":"ContainerDied","Data":"b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1"} Dec 07 09:29:19 crc kubenswrapper[4838]: I1207 09:29:19.498212 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-lv5g7" event={"ID":"2eafdf67-be05-41b8-b67a-7ba69fd9774f","Type":"ContainerStarted","Data":"4ea9a6bf303d83e20616e4f9d91fecfa256b21bba6fbb84d2b7222f079d35a09"} Dec 07 09:29:20 crc kubenswrapper[4838]: I1207 09:29:20.509099 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-lv5g7" event={"ID":"2eafdf67-be05-41b8-b67a-7ba69fd9774f","Type":"ContainerStarted","Data":"afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241"} Dec 07 09:29:20 crc kubenswrapper[4838]: I1207 09:29:20.509581 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:20 crc kubenswrapper[4838]: I1207 09:29:20.542943 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8676544467-lv5g7" podStartSLOduration=3.542923724 podStartE2EDuration="3.542923724s" podCreationTimestamp="2025-12-07 09:29:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:29:20.526965544 +0000 UTC m=+1377.234284561" watchObservedRunningTime="2025-12-07 09:29:20.542923724 +0000 UTC m=+1377.250242751" Dec 07 09:29:21 crc kubenswrapper[4838]: I1207 09:29:21.518030 4838 generic.go:334] "Generic (PLEG): container finished" podID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerID="080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128" exitCode=0 Dec 07 09:29:21 crc kubenswrapper[4838]: I1207 09:29:21.518099 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s96k" event={"ID":"05544f2f-de2e-4a39-9b19-79542a4d3508","Type":"ContainerDied","Data":"080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128"} Dec 07 09:29:22 crc kubenswrapper[4838]: I1207 09:29:22.527389 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s96k" event={"ID":"05544f2f-de2e-4a39-9b19-79542a4d3508","Type":"ContainerStarted","Data":"e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab"} Dec 07 09:29:22 crc kubenswrapper[4838]: I1207 09:29:22.552535 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9s96k" podStartSLOduration=2.015394136 podStartE2EDuration="6.552511852s" podCreationTimestamp="2025-12-07 09:29:16 +0000 UTC" firstStartedPulling="2025-12-07 09:29:17.477526911 +0000 UTC m=+1374.184845938" lastFinishedPulling="2025-12-07 09:29:22.014644637 +0000 UTC m=+1378.721963654" observedRunningTime="2025-12-07 09:29:22.544540002 +0000 UTC m=+1379.251859029" watchObservedRunningTime="2025-12-07 09:29:22.552511852 +0000 UTC m=+1379.259830869" Dec 07 09:29:26 crc kubenswrapper[4838]: I1207 09:29:26.732792 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:26 crc kubenswrapper[4838]: I1207 09:29:26.733331 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:27 crc kubenswrapper[4838]: I1207 09:29:27.795348 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9s96k" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="registry-server" probeResult="failure" output=< Dec 07 09:29:27 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 09:29:27 crc kubenswrapper[4838]: > Dec 07 09:29:27 crc kubenswrapper[4838]: I1207 09:29:27.934064 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.006069 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-lsr6s"] Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.006361 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerName="dnsmasq-dns" containerID="cri-o://4dc72005e29e40e16a332f0c82e3ea5dec6046c64ef78695f7f8b1cda5882398" gracePeriod=10 Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.197733 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8454874f4c-gkmgf"] Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.205101 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.225946 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8454874f4c-gkmgf"] Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.322340 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt5wm\" (UniqueName: \"kubernetes.io/projected/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-kube-api-access-xt5wm\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.322453 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-openstack-edpm-ipam\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.322482 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-sb\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.322528 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-config\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.322552 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-dns-svc\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.322578 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-nb\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.424672 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-nb\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.424732 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt5wm\" (UniqueName: \"kubernetes.io/projected/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-kube-api-access-xt5wm\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.424824 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-openstack-edpm-ipam\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.424855 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-sb\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.424887 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-config\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.424909 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-dns-svc\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.426497 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-dns-svc\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.427154 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-sb\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.427291 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-config\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.427923 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-openstack-edpm-ipam\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.428560 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-nb\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.452774 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt5wm\" (UniqueName: \"kubernetes.io/projected/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-kube-api-access-xt5wm\") pod \"dnsmasq-dns-8454874f4c-gkmgf\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.587665 4838 generic.go:334] "Generic (PLEG): container finished" podID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerID="4dc72005e29e40e16a332f0c82e3ea5dec6046c64ef78695f7f8b1cda5882398" exitCode=0 Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.587707 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" event={"ID":"45c647ef-5d0b-44ea-96f9-8c4c00c237f9","Type":"ContainerDied","Data":"4dc72005e29e40e16a332f0c82e3ea5dec6046c64ef78695f7f8b1cda5882398"} Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.587734 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" event={"ID":"45c647ef-5d0b-44ea-96f9-8c4c00c237f9","Type":"ContainerDied","Data":"c83338f81916872e66027b04df0bada58447605a1f124fcee72bd5585ddc0830"} Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.587769 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c83338f81916872e66027b04df0bada58447605a1f124fcee72bd5585ddc0830" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.588194 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.598208 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.734287 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-config\") pod \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.735329 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-dns-svc\") pod \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.735349 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-nb\") pod \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.735373 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-sb\") pod \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.735433 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgk4g\" (UniqueName: \"kubernetes.io/projected/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-kube-api-access-rgk4g\") pod \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\" (UID: \"45c647ef-5d0b-44ea-96f9-8c4c00c237f9\") " Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.763930 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-kube-api-access-rgk4g" (OuterVolumeSpecName: "kube-api-access-rgk4g") pod "45c647ef-5d0b-44ea-96f9-8c4c00c237f9" (UID: "45c647ef-5d0b-44ea-96f9-8c4c00c237f9"). InnerVolumeSpecName "kube-api-access-rgk4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.828916 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "45c647ef-5d0b-44ea-96f9-8c4c00c237f9" (UID: "45c647ef-5d0b-44ea-96f9-8c4c00c237f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.838146 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.838172 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgk4g\" (UniqueName: \"kubernetes.io/projected/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-kube-api-access-rgk4g\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.864445 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "45c647ef-5d0b-44ea-96f9-8c4c00c237f9" (UID: "45c647ef-5d0b-44ea-96f9-8c4c00c237f9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.879309 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "45c647ef-5d0b-44ea-96f9-8c4c00c237f9" (UID: "45c647ef-5d0b-44ea-96f9-8c4c00c237f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.907262 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-config" (OuterVolumeSpecName: "config") pod "45c647ef-5d0b-44ea-96f9-8c4c00c237f9" (UID: "45c647ef-5d0b-44ea-96f9-8c4c00c237f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.941949 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.941984 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:28 crc kubenswrapper[4838]: I1207 09:29:28.941993 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45c647ef-5d0b-44ea-96f9-8c4c00c237f9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:29 crc kubenswrapper[4838]: I1207 09:29:29.058796 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8454874f4c-gkmgf"] Dec 07 09:29:29 crc kubenswrapper[4838]: W1207 09:29:29.065012 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8db08fda_8fbf_441f_ac6e_ba4e76ed9480.slice/crio-12e559f44177754e3d65f30f9147467512a6b5c5119aa99d725c253c5cceb970 WatchSource:0}: Error finding container 12e559f44177754e3d65f30f9147467512a6b5c5119aa99d725c253c5cceb970: Status 404 returned error can't find the container with id 12e559f44177754e3d65f30f9147467512a6b5c5119aa99d725c253c5cceb970 Dec 07 09:29:29 crc kubenswrapper[4838]: I1207 09:29:29.599200 4838 generic.go:334] "Generic (PLEG): container finished" podID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerID="851a9e33e15edd365080edb35f8628de761d087302c83c79ad8eb2d99a020a63" exitCode=0 Dec 07 09:29:29 crc kubenswrapper[4838]: I1207 09:29:29.599299 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" event={"ID":"8db08fda-8fbf-441f-ac6e-ba4e76ed9480","Type":"ContainerDied","Data":"851a9e33e15edd365080edb35f8628de761d087302c83c79ad8eb2d99a020a63"} Dec 07 09:29:29 crc kubenswrapper[4838]: I1207 09:29:29.599367 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" event={"ID":"8db08fda-8fbf-441f-ac6e-ba4e76ed9480","Type":"ContainerStarted","Data":"12e559f44177754e3d65f30f9147467512a6b5c5119aa99d725c253c5cceb970"} Dec 07 09:29:29 crc kubenswrapper[4838]: I1207 09:29:29.599336 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:29:30 crc kubenswrapper[4838]: I1207 09:29:30.609540 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" event={"ID":"8db08fda-8fbf-441f-ac6e-ba4e76ed9480","Type":"ContainerStarted","Data":"0b3c9353597d233272b320564e6a4203dcd87e3522d5138d7574877b61365b88"} Dec 07 09:29:30 crc kubenswrapper[4838]: I1207 09:29:30.609919 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:30 crc kubenswrapper[4838]: I1207 09:29:30.636086 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" podStartSLOduration=2.6359175219999997 podStartE2EDuration="2.635917522s" podCreationTimestamp="2025-12-07 09:29:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:29:30.633091121 +0000 UTC m=+1387.340410148" watchObservedRunningTime="2025-12-07 09:29:30.635917522 +0000 UTC m=+1387.343236559" Dec 07 09:29:36 crc kubenswrapper[4838]: I1207 09:29:36.779327 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:36 crc kubenswrapper[4838]: I1207 09:29:36.828982 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:37 crc kubenswrapper[4838]: I1207 09:29:37.020301 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9s96k"] Dec 07 09:29:38 crc kubenswrapper[4838]: I1207 09:29:38.598055 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 09:29:38 crc kubenswrapper[4838]: I1207 09:29:38.688328 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8676544467-lv5g7"] Dec 07 09:29:38 crc kubenswrapper[4838]: I1207 09:29:38.689859 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8676544467-lv5g7" podUID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerName="dnsmasq-dns" containerID="cri-o://afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241" gracePeriod=10 Dec 07 09:29:38 crc kubenswrapper[4838]: I1207 09:29:38.703261 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9s96k" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="registry-server" containerID="cri-o://e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab" gracePeriod=2 Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.129851 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.200598 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222335 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-dns-svc\") pod \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222438 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gr9s\" (UniqueName: \"kubernetes.io/projected/05544f2f-de2e-4a39-9b19-79542a4d3508-kube-api-access-6gr9s\") pod \"05544f2f-de2e-4a39-9b19-79542a4d3508\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222538 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-config\") pod \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222597 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-catalog-content\") pod \"05544f2f-de2e-4a39-9b19-79542a4d3508\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222637 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-nb\") pod \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222715 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-sb\") pod \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222776 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-utilities\") pod \"05544f2f-de2e-4a39-9b19-79542a4d3508\" (UID: \"05544f2f-de2e-4a39-9b19-79542a4d3508\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222880 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-openstack-edpm-ipam\") pod \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.222927 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4vtz\" (UniqueName: \"kubernetes.io/projected/2eafdf67-be05-41b8-b67a-7ba69fd9774f-kube-api-access-b4vtz\") pod \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\" (UID: \"2eafdf67-be05-41b8-b67a-7ba69fd9774f\") " Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.224988 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-utilities" (OuterVolumeSpecName: "utilities") pod "05544f2f-de2e-4a39-9b19-79542a4d3508" (UID: "05544f2f-de2e-4a39-9b19-79542a4d3508"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.231072 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eafdf67-be05-41b8-b67a-7ba69fd9774f-kube-api-access-b4vtz" (OuterVolumeSpecName: "kube-api-access-b4vtz") pod "2eafdf67-be05-41b8-b67a-7ba69fd9774f" (UID: "2eafdf67-be05-41b8-b67a-7ba69fd9774f"). InnerVolumeSpecName "kube-api-access-b4vtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.248262 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05544f2f-de2e-4a39-9b19-79542a4d3508-kube-api-access-6gr9s" (OuterVolumeSpecName: "kube-api-access-6gr9s") pod "05544f2f-de2e-4a39-9b19-79542a4d3508" (UID: "05544f2f-de2e-4a39-9b19-79542a4d3508"). InnerVolumeSpecName "kube-api-access-6gr9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.274137 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-config" (OuterVolumeSpecName: "config") pod "2eafdf67-be05-41b8-b67a-7ba69fd9774f" (UID: "2eafdf67-be05-41b8-b67a-7ba69fd9774f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.301927 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2eafdf67-be05-41b8-b67a-7ba69fd9774f" (UID: "2eafdf67-be05-41b8-b67a-7ba69fd9774f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.304212 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2eafdf67-be05-41b8-b67a-7ba69fd9774f" (UID: "2eafdf67-be05-41b8-b67a-7ba69fd9774f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.319664 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2eafdf67-be05-41b8-b67a-7ba69fd9774f" (UID: "2eafdf67-be05-41b8-b67a-7ba69fd9774f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.325810 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.325872 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4vtz\" (UniqueName: \"kubernetes.io/projected/2eafdf67-be05-41b8-b67a-7ba69fd9774f-kube-api-access-b4vtz\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.325883 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.325892 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gr9s\" (UniqueName: \"kubernetes.io/projected/05544f2f-de2e-4a39-9b19-79542a4d3508-kube-api-access-6gr9s\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.325901 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-config\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.325910 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.325918 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.327297 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "2eafdf67-be05-41b8-b67a-7ba69fd9774f" (UID: "2eafdf67-be05-41b8-b67a-7ba69fd9774f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.380053 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05544f2f-de2e-4a39-9b19-79542a4d3508" (UID: "05544f2f-de2e-4a39-9b19-79542a4d3508"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.426855 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05544f2f-de2e-4a39-9b19-79542a4d3508-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.426890 4838 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2eafdf67-be05-41b8-b67a-7ba69fd9774f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.713248 4838 generic.go:334] "Generic (PLEG): container finished" podID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerID="afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241" exitCode=0 Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.713352 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-lv5g7" event={"ID":"2eafdf67-be05-41b8-b67a-7ba69fd9774f","Type":"ContainerDied","Data":"afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241"} Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.713375 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-lv5g7" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.713424 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-lv5g7" event={"ID":"2eafdf67-be05-41b8-b67a-7ba69fd9774f","Type":"ContainerDied","Data":"4ea9a6bf303d83e20616e4f9d91fecfa256b21bba6fbb84d2b7222f079d35a09"} Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.713446 4838 scope.go:117] "RemoveContainer" containerID="afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.721745 4838 generic.go:334] "Generic (PLEG): container finished" podID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerID="e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab" exitCode=0 Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.721848 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s96k" event={"ID":"05544f2f-de2e-4a39-9b19-79542a4d3508","Type":"ContainerDied","Data":"e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab"} Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.721891 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9s96k" event={"ID":"05544f2f-de2e-4a39-9b19-79542a4d3508","Type":"ContainerDied","Data":"2d1d6189ee79152e14ddc8787b793251d24f0611f79cd08a69cd748a67dd8255"} Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.722115 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9s96k" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.752451 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8676544467-lv5g7"] Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.757294 4838 scope.go:117] "RemoveContainer" containerID="b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.780992 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8676544467-lv5g7"] Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.790216 4838 scope.go:117] "RemoveContainer" containerID="afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241" Dec 07 09:29:39 crc kubenswrapper[4838]: E1207 09:29:39.790903 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241\": container with ID starting with afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241 not found: ID does not exist" containerID="afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.790956 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241"} err="failed to get container status \"afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241\": rpc error: code = NotFound desc = could not find container \"afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241\": container with ID starting with afc3b45a4e9ca4d22231f32b61bc111bb29b268a007211109cfb24f7fbe9f241 not found: ID does not exist" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.790987 4838 scope.go:117] "RemoveContainer" containerID="b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1" Dec 07 09:29:39 crc kubenswrapper[4838]: E1207 09:29:39.791302 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1\": container with ID starting with b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1 not found: ID does not exist" containerID="b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.791332 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1"} err="failed to get container status \"b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1\": rpc error: code = NotFound desc = could not find container \"b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1\": container with ID starting with b990c052a6ab3cfece1645ad8ca34909410cf64a2e94328647e7dbc2588445b1 not found: ID does not exist" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.791361 4838 scope.go:117] "RemoveContainer" containerID="e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.801003 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9s96k"] Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.809002 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9s96k"] Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.877882 4838 scope.go:117] "RemoveContainer" containerID="080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.905923 4838 scope.go:117] "RemoveContainer" containerID="e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.955260 4838 scope.go:117] "RemoveContainer" containerID="e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab" Dec 07 09:29:39 crc kubenswrapper[4838]: E1207 09:29:39.956260 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab\": container with ID starting with e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab not found: ID does not exist" containerID="e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.956297 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab"} err="failed to get container status \"e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab\": rpc error: code = NotFound desc = could not find container \"e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab\": container with ID starting with e3c5d9f9b5d54667e8e76e642b345c4b1f0b5b3e4df13bba8d1756ef93cccbab not found: ID does not exist" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.956326 4838 scope.go:117] "RemoveContainer" containerID="080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128" Dec 07 09:29:39 crc kubenswrapper[4838]: E1207 09:29:39.957026 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128\": container with ID starting with 080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128 not found: ID does not exist" containerID="080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.957053 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128"} err="failed to get container status \"080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128\": rpc error: code = NotFound desc = could not find container \"080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128\": container with ID starting with 080a4a6001f1083be399bb776330d6a2a0a18302a3e410bd41caa88ea6574128 not found: ID does not exist" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.957070 4838 scope.go:117] "RemoveContainer" containerID="e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5" Dec 07 09:29:39 crc kubenswrapper[4838]: E1207 09:29:39.957459 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5\": container with ID starting with e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5 not found: ID does not exist" containerID="e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5" Dec 07 09:29:39 crc kubenswrapper[4838]: I1207 09:29:39.957483 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5"} err="failed to get container status \"e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5\": rpc error: code = NotFound desc = could not find container \"e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5\": container with ID starting with e6af96e2e915eb05552273ee686844d328846375e239b07339344c74650b7ef5 not found: ID does not exist" Dec 07 09:29:41 crc kubenswrapper[4838]: I1207 09:29:41.635030 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" path="/var/lib/kubelet/pods/05544f2f-de2e-4a39-9b19-79542a4d3508/volumes" Dec 07 09:29:41 crc kubenswrapper[4838]: I1207 09:29:41.636279 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" path="/var/lib/kubelet/pods/2eafdf67-be05-41b8-b67a-7ba69fd9774f/volumes" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.788068 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484"] Dec 07 09:29:48 crc kubenswrapper[4838]: E1207 09:29:48.790527 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="registry-server" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.790615 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="registry-server" Dec 07 09:29:48 crc kubenswrapper[4838]: E1207 09:29:48.790684 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="extract-content" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.790779 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="extract-content" Dec 07 09:29:48 crc kubenswrapper[4838]: E1207 09:29:48.790948 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerName="init" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.791008 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerName="init" Dec 07 09:29:48 crc kubenswrapper[4838]: E1207 09:29:48.791076 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerName="dnsmasq-dns" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.791130 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerName="dnsmasq-dns" Dec 07 09:29:48 crc kubenswrapper[4838]: E1207 09:29:48.791186 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerName="init" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.791242 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerName="init" Dec 07 09:29:48 crc kubenswrapper[4838]: E1207 09:29:48.791308 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerName="dnsmasq-dns" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.791361 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerName="dnsmasq-dns" Dec 07 09:29:48 crc kubenswrapper[4838]: E1207 09:29:48.791416 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="extract-utilities" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.791477 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="extract-utilities" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.791693 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" containerName="dnsmasq-dns" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.791778 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eafdf67-be05-41b8-b67a-7ba69fd9774f" containerName="dnsmasq-dns" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.793639 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="05544f2f-de2e-4a39-9b19-79542a4d3508" containerName="registry-server" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.794659 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.796883 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.797070 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.797515 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.798351 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.807425 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484"] Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.819483 4838 generic.go:334] "Generic (PLEG): container finished" podID="688c3086-fd09-4b9d-921e-e41c7f8a01ac" containerID="8f03bef63131f4e707fc2572584aace63debb4cd682deeabffb08f14d1f33db7" exitCode=0 Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.819529 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"688c3086-fd09-4b9d-921e-e41c7f8a01ac","Type":"ContainerDied","Data":"8f03bef63131f4e707fc2572584aace63debb4cd682deeabffb08f14d1f33db7"} Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.910556 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tjgx\" (UniqueName: \"kubernetes.io/projected/333d9a84-4e02-4851-b612-0cf2b6677656-kube-api-access-6tjgx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.910804 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.911008 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:48 crc kubenswrapper[4838]: I1207 09:29:48.911139 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.012907 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.013502 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.013560 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tjgx\" (UniqueName: \"kubernetes.io/projected/333d9a84-4e02-4851-b612-0cf2b6677656-kube-api-access-6tjgx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.013587 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.020408 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.021054 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.021714 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.050367 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tjgx\" (UniqueName: \"kubernetes.io/projected/333d9a84-4e02-4851-b612-0cf2b6677656-kube-api-access-6tjgx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k4484\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.115229 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:29:49 crc kubenswrapper[4838]: W1207 09:29:49.522350 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333d9a84_4e02_4851_b612_0cf2b6677656.slice/crio-a5937f3febde28487c6283b4a7cc121eb90e75e7fc0db370d41e9db98dd645f2 WatchSource:0}: Error finding container a5937f3febde28487c6283b4a7cc121eb90e75e7fc0db370d41e9db98dd645f2: Status 404 returned error can't find the container with id a5937f3febde28487c6283b4a7cc121eb90e75e7fc0db370d41e9db98dd645f2 Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.527412 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484"] Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.829223 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" event={"ID":"333d9a84-4e02-4851-b612-0cf2b6677656","Type":"ContainerStarted","Data":"a5937f3febde28487c6283b4a7cc121eb90e75e7fc0db370d41e9db98dd645f2"} Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.830959 4838 generic.go:334] "Generic (PLEG): container finished" podID="a017a788-cb2e-45ed-bceb-26ea6b69a02d" containerID="65bddec28c029221a0a8a86d4939de10e2a37675e5f15f2e6a80fb125b3ebce6" exitCode=0 Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.831027 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a017a788-cb2e-45ed-bceb-26ea6b69a02d","Type":"ContainerDied","Data":"65bddec28c029221a0a8a86d4939de10e2a37675e5f15f2e6a80fb125b3ebce6"} Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.835922 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"688c3086-fd09-4b9d-921e-e41c7f8a01ac","Type":"ContainerStarted","Data":"aaed127049a94dbd020e2cf357ea86eaf04d7013892a4d55ff3b63e3c534d8b9"} Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.836246 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 07 09:29:49 crc kubenswrapper[4838]: I1207 09:29:49.904007 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.903983432 podStartE2EDuration="36.903983432s" podCreationTimestamp="2025-12-07 09:29:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:29:49.890888744 +0000 UTC m=+1406.598207781" watchObservedRunningTime="2025-12-07 09:29:49.903983432 +0000 UTC m=+1406.611302449" Dec 07 09:29:50 crc kubenswrapper[4838]: I1207 09:29:50.847003 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a017a788-cb2e-45ed-bceb-26ea6b69a02d","Type":"ContainerStarted","Data":"8a4235dab77ec5005e0adf43e3af2da7c19af7f1d85de2a1bcd8f31c28019a91"} Dec 07 09:29:50 crc kubenswrapper[4838]: I1207 09:29:50.847348 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:29:50 crc kubenswrapper[4838]: I1207 09:29:50.869066 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.86904647 podStartE2EDuration="35.86904647s" podCreationTimestamp="2025-12-07 09:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 09:29:50.866678291 +0000 UTC m=+1407.573997308" watchObservedRunningTime="2025-12-07 09:29:50.86904647 +0000 UTC m=+1407.576365487" Dec 07 09:29:59 crc kubenswrapper[4838]: I1207 09:29:59.754092 4838 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod45c647ef-5d0b-44ea-96f9-8c4c00c237f9"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod45c647ef-5d0b-44ea-96f9-8c4c00c237f9] : Timed out while waiting for systemd to remove kubepods-besteffort-pod45c647ef_5d0b_44ea_96f9_8c4c00c237f9.slice" Dec 07 09:29:59 crc kubenswrapper[4838]: E1207 09:29:59.754625 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod45c647ef-5d0b-44ea-96f9-8c4c00c237f9] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod45c647ef-5d0b-44ea-96f9-8c4c00c237f9] : Timed out while waiting for systemd to remove kubepods-besteffort-pod45c647ef_5d0b_44ea_96f9_8c4c00c237f9.slice" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" Dec 07 09:29:59 crc kubenswrapper[4838]: I1207 09:29:59.941232 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-lsr6s" Dec 07 09:29:59 crc kubenswrapper[4838]: I1207 09:29:59.979918 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-lsr6s"] Dec 07 09:29:59 crc kubenswrapper[4838]: I1207 09:29:59.986676 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-lsr6s"] Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.153027 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t"] Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.154435 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.157565 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.163049 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t"] Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.164310 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.315180 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6glnj\" (UniqueName: \"kubernetes.io/projected/7e88c0c3-c7da-469e-b34e-7e2d7672611c-kube-api-access-6glnj\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.315276 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e88c0c3-c7da-469e-b34e-7e2d7672611c-secret-volume\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.315315 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e88c0c3-c7da-469e-b34e-7e2d7672611c-config-volume\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.417238 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e88c0c3-c7da-469e-b34e-7e2d7672611c-secret-volume\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.417299 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e88c0c3-c7da-469e-b34e-7e2d7672611c-config-volume\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.417392 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6glnj\" (UniqueName: \"kubernetes.io/projected/7e88c0c3-c7da-469e-b34e-7e2d7672611c-kube-api-access-6glnj\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.418376 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e88c0c3-c7da-469e-b34e-7e2d7672611c-config-volume\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.439272 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e88c0c3-c7da-469e-b34e-7e2d7672611c-secret-volume\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.441579 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6glnj\" (UniqueName: \"kubernetes.io/projected/7e88c0c3-c7da-469e-b34e-7e2d7672611c-kube-api-access-6glnj\") pod \"collect-profiles-29418330-gbw9t\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:00 crc kubenswrapper[4838]: I1207 09:30:00.475920 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:01 crc kubenswrapper[4838]: I1207 09:30:01.634465 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45c647ef-5d0b-44ea-96f9-8c4c00c237f9" path="/var/lib/kubelet/pods/45c647ef-5d0b-44ea-96f9-8c4c00c237f9/volumes" Dec 07 09:30:04 crc kubenswrapper[4838]: I1207 09:30:04.067070 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 07 09:30:04 crc kubenswrapper[4838]: I1207 09:30:04.283553 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t"] Dec 07 09:30:05 crc kubenswrapper[4838]: I1207 09:30:05.015983 4838 generic.go:334] "Generic (PLEG): container finished" podID="7e88c0c3-c7da-469e-b34e-7e2d7672611c" containerID="a8d433a11c4db30548a4e54a6a8e6119f2415d8689c13931367e21a2c50e97bd" exitCode=0 Dec 07 09:30:05 crc kubenswrapper[4838]: I1207 09:30:05.016096 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" event={"ID":"7e88c0c3-c7da-469e-b34e-7e2d7672611c","Type":"ContainerDied","Data":"a8d433a11c4db30548a4e54a6a8e6119f2415d8689c13931367e21a2c50e97bd"} Dec 07 09:30:05 crc kubenswrapper[4838]: I1207 09:30:05.016358 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" event={"ID":"7e88c0c3-c7da-469e-b34e-7e2d7672611c","Type":"ContainerStarted","Data":"88021eb373f61b6ae4e9e147b6b3592551d0948df6926c84aaa4e6d7fb808581"} Dec 07 09:30:05 crc kubenswrapper[4838]: I1207 09:30:05.018268 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" event={"ID":"333d9a84-4e02-4851-b612-0cf2b6677656","Type":"ContainerStarted","Data":"967cd965ac848b511b7edf75c70dcfdb50b08a6b2bbab12248ae3712810291c8"} Dec 07 09:30:05 crc kubenswrapper[4838]: I1207 09:30:05.060509 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" podStartSLOduration=2.694444674 podStartE2EDuration="17.060488531s" podCreationTimestamp="2025-12-07 09:29:48 +0000 UTC" firstStartedPulling="2025-12-07 09:29:49.524797354 +0000 UTC m=+1406.232116371" lastFinishedPulling="2025-12-07 09:30:03.890841211 +0000 UTC m=+1420.598160228" observedRunningTime="2025-12-07 09:30:05.059725679 +0000 UTC m=+1421.767044696" watchObservedRunningTime="2025-12-07 09:30:05.060488531 +0000 UTC m=+1421.767807548" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.127524 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.462767 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.539648 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e88c0c3-c7da-469e-b34e-7e2d7672611c-config-volume\") pod \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.539723 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6glnj\" (UniqueName: \"kubernetes.io/projected/7e88c0c3-c7da-469e-b34e-7e2d7672611c-kube-api-access-6glnj\") pod \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.539750 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e88c0c3-c7da-469e-b34e-7e2d7672611c-secret-volume\") pod \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\" (UID: \"7e88c0c3-c7da-469e-b34e-7e2d7672611c\") " Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.541463 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e88c0c3-c7da-469e-b34e-7e2d7672611c-config-volume" (OuterVolumeSpecName: "config-volume") pod "7e88c0c3-c7da-469e-b34e-7e2d7672611c" (UID: "7e88c0c3-c7da-469e-b34e-7e2d7672611c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.547891 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e88c0c3-c7da-469e-b34e-7e2d7672611c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7e88c0c3-c7da-469e-b34e-7e2d7672611c" (UID: "7e88c0c3-c7da-469e-b34e-7e2d7672611c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.554925 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e88c0c3-c7da-469e-b34e-7e2d7672611c-kube-api-access-6glnj" (OuterVolumeSpecName: "kube-api-access-6glnj") pod "7e88c0c3-c7da-469e-b34e-7e2d7672611c" (UID: "7e88c0c3-c7da-469e-b34e-7e2d7672611c"). InnerVolumeSpecName "kube-api-access-6glnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.641312 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e88c0c3-c7da-469e-b34e-7e2d7672611c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.641342 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6glnj\" (UniqueName: \"kubernetes.io/projected/7e88c0c3-c7da-469e-b34e-7e2d7672611c-kube-api-access-6glnj\") on node \"crc\" DevicePath \"\"" Dec 07 09:30:06 crc kubenswrapper[4838]: I1207 09:30:06.641355 4838 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e88c0c3-c7da-469e-b34e-7e2d7672611c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:30:07 crc kubenswrapper[4838]: I1207 09:30:07.038342 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" event={"ID":"7e88c0c3-c7da-469e-b34e-7e2d7672611c","Type":"ContainerDied","Data":"88021eb373f61b6ae4e9e147b6b3592551d0948df6926c84aaa4e6d7fb808581"} Dec 07 09:30:07 crc kubenswrapper[4838]: I1207 09:30:07.038675 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88021eb373f61b6ae4e9e147b6b3592551d0948df6926c84aaa4e6d7fb808581" Dec 07 09:30:07 crc kubenswrapper[4838]: I1207 09:30:07.038421 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t" Dec 07 09:30:16 crc kubenswrapper[4838]: I1207 09:30:16.111005 4838 generic.go:334] "Generic (PLEG): container finished" podID="333d9a84-4e02-4851-b612-0cf2b6677656" containerID="967cd965ac848b511b7edf75c70dcfdb50b08a6b2bbab12248ae3712810291c8" exitCode=0 Dec 07 09:30:16 crc kubenswrapper[4838]: I1207 09:30:16.111034 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" event={"ID":"333d9a84-4e02-4851-b612-0cf2b6677656","Type":"ContainerDied","Data":"967cd965ac848b511b7edf75c70dcfdb50b08a6b2bbab12248ae3712810291c8"} Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.646118 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.671991 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-inventory\") pod \"333d9a84-4e02-4851-b612-0cf2b6677656\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.672052 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tjgx\" (UniqueName: \"kubernetes.io/projected/333d9a84-4e02-4851-b612-0cf2b6677656-kube-api-access-6tjgx\") pod \"333d9a84-4e02-4851-b612-0cf2b6677656\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.672144 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-ssh-key\") pod \"333d9a84-4e02-4851-b612-0cf2b6677656\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.672202 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-repo-setup-combined-ca-bundle\") pod \"333d9a84-4e02-4851-b612-0cf2b6677656\" (UID: \"333d9a84-4e02-4851-b612-0cf2b6677656\") " Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.681045 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333d9a84-4e02-4851-b612-0cf2b6677656-kube-api-access-6tjgx" (OuterVolumeSpecName: "kube-api-access-6tjgx") pod "333d9a84-4e02-4851-b612-0cf2b6677656" (UID: "333d9a84-4e02-4851-b612-0cf2b6677656"). InnerVolumeSpecName "kube-api-access-6tjgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.699029 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "333d9a84-4e02-4851-b612-0cf2b6677656" (UID: "333d9a84-4e02-4851-b612-0cf2b6677656"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.764974 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "333d9a84-4e02-4851-b612-0cf2b6677656" (UID: "333d9a84-4e02-4851-b612-0cf2b6677656"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.777720 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tjgx\" (UniqueName: \"kubernetes.io/projected/333d9a84-4e02-4851-b612-0cf2b6677656-kube-api-access-6tjgx\") on node \"crc\" DevicePath \"\"" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.777970 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.778034 4838 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.802009 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-inventory" (OuterVolumeSpecName: "inventory") pod "333d9a84-4e02-4851-b612-0cf2b6677656" (UID: "333d9a84-4e02-4851-b612-0cf2b6677656"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:30:17 crc kubenswrapper[4838]: I1207 09:30:17.878966 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/333d9a84-4e02-4851-b612-0cf2b6677656-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.135157 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" event={"ID":"333d9a84-4e02-4851-b612-0cf2b6677656","Type":"ContainerDied","Data":"a5937f3febde28487c6283b4a7cc121eb90e75e7fc0db370d41e9db98dd645f2"} Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.135205 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5937f3febde28487c6283b4a7cc121eb90e75e7fc0db370d41e9db98dd645f2" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.135290 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.228283 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82"] Dec 07 09:30:18 crc kubenswrapper[4838]: E1207 09:30:18.228656 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333d9a84-4e02-4851-b612-0cf2b6677656" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.228673 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="333d9a84-4e02-4851-b612-0cf2b6677656" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 09:30:18 crc kubenswrapper[4838]: E1207 09:30:18.228708 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e88c0c3-c7da-469e-b34e-7e2d7672611c" containerName="collect-profiles" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.228714 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e88c0c3-c7da-469e-b34e-7e2d7672611c" containerName="collect-profiles" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.228923 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e88c0c3-c7da-469e-b34e-7e2d7672611c" containerName="collect-profiles" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.228942 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="333d9a84-4e02-4851-b612-0cf2b6677656" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.229505 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.231694 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.232566 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.240837 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.241167 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.255484 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82"] Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.286362 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.286486 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.286519 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.286576 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx9q2\" (UniqueName: \"kubernetes.io/projected/9e993146-3e36-443e-b88b-1d090aa17900-kube-api-access-sx9q2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.387867 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.388132 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx9q2\" (UniqueName: \"kubernetes.io/projected/9e993146-3e36-443e-b88b-1d090aa17900-kube-api-access-sx9q2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.388194 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.388257 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.391555 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.392145 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.397356 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.403535 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx9q2\" (UniqueName: \"kubernetes.io/projected/9e993146-3e36-443e-b88b-1d090aa17900-kube-api-access-sx9q2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:18 crc kubenswrapper[4838]: I1207 09:30:18.576142 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:30:19 crc kubenswrapper[4838]: I1207 09:30:19.093262 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82"] Dec 07 09:30:19 crc kubenswrapper[4838]: I1207 09:30:19.143301 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" event={"ID":"9e993146-3e36-443e-b88b-1d090aa17900","Type":"ContainerStarted","Data":"64bda83a9f9966223defb6b92ca8f2bcbce472e4684c7cbbe7d83c0f87580e02"} Dec 07 09:30:20 crc kubenswrapper[4838]: I1207 09:30:20.154301 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" event={"ID":"9e993146-3e36-443e-b88b-1d090aa17900","Type":"ContainerStarted","Data":"19d85c270699df25aee8183fa0de7b595a3af3a8a51966a213675c01c342beb3"} Dec 07 09:30:20 crc kubenswrapper[4838]: I1207 09:30:20.179383 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" podStartSLOduration=1.7997095029999999 podStartE2EDuration="2.179360697s" podCreationTimestamp="2025-12-07 09:30:18 +0000 UTC" firstStartedPulling="2025-12-07 09:30:19.093010144 +0000 UTC m=+1435.800329161" lastFinishedPulling="2025-12-07 09:30:19.472661338 +0000 UTC m=+1436.179980355" observedRunningTime="2025-12-07 09:30:20.178584574 +0000 UTC m=+1436.885903591" watchObservedRunningTime="2025-12-07 09:30:20.179360697 +0000 UTC m=+1436.886679724" Dec 07 09:30:49 crc kubenswrapper[4838]: I1207 09:30:49.392928 4838 scope.go:117] "RemoveContainer" containerID="5eb7b29805ce8fb4b7990c6b189e9f9990dada903daeac5e7f823d8a21266dbe" Dec 07 09:30:49 crc kubenswrapper[4838]: I1207 09:30:49.421493 4838 scope.go:117] "RemoveContainer" containerID="9e50b3fe4418a1ee0920041fb831a5a2ad19591f0c794ec8b49af11dbed1720b" Dec 07 09:30:54 crc kubenswrapper[4838]: I1207 09:30:54.493240 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:30:54 crc kubenswrapper[4838]: I1207 09:30:54.493754 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:31:24 crc kubenswrapper[4838]: I1207 09:31:24.493733 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:31:24 crc kubenswrapper[4838]: I1207 09:31:24.494320 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.770143 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vps95"] Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.776532 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.801431 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vps95"] Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.880795 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-catalog-content\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.880914 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhltd\" (UniqueName: \"kubernetes.io/projected/b2ac318d-94b4-4ba4-8944-d597ce82db2e-kube-api-access-lhltd\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.880951 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-utilities\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.982851 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-catalog-content\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.982936 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhltd\" (UniqueName: \"kubernetes.io/projected/b2ac318d-94b4-4ba4-8944-d597ce82db2e-kube-api-access-lhltd\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.982972 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-utilities\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.983432 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-utilities\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:33 crc kubenswrapper[4838]: I1207 09:31:33.983529 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-catalog-content\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:34 crc kubenswrapper[4838]: I1207 09:31:34.021281 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhltd\" (UniqueName: \"kubernetes.io/projected/b2ac318d-94b4-4ba4-8944-d597ce82db2e-kube-api-access-lhltd\") pod \"certified-operators-vps95\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:34 crc kubenswrapper[4838]: I1207 09:31:34.102185 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:34 crc kubenswrapper[4838]: I1207 09:31:34.695698 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vps95"] Dec 07 09:31:34 crc kubenswrapper[4838]: I1207 09:31:34.935625 4838 generic.go:334] "Generic (PLEG): container finished" podID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerID="df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82" exitCode=0 Dec 07 09:31:34 crc kubenswrapper[4838]: I1207 09:31:34.935661 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vps95" event={"ID":"b2ac318d-94b4-4ba4-8944-d597ce82db2e","Type":"ContainerDied","Data":"df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82"} Dec 07 09:31:34 crc kubenswrapper[4838]: I1207 09:31:34.935934 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vps95" event={"ID":"b2ac318d-94b4-4ba4-8944-d597ce82db2e","Type":"ContainerStarted","Data":"2da47d3267cfb02bbeeb6148c6ee7c914e2e1165da79d00c8255ff61d8f14f9b"} Dec 07 09:31:35 crc kubenswrapper[4838]: I1207 09:31:35.945948 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vps95" event={"ID":"b2ac318d-94b4-4ba4-8944-d597ce82db2e","Type":"ContainerStarted","Data":"82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7"} Dec 07 09:31:37 crc kubenswrapper[4838]: I1207 09:31:37.963272 4838 generic.go:334] "Generic (PLEG): container finished" podID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerID="82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7" exitCode=0 Dec 07 09:31:37 crc kubenswrapper[4838]: I1207 09:31:37.963590 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vps95" event={"ID":"b2ac318d-94b4-4ba4-8944-d597ce82db2e","Type":"ContainerDied","Data":"82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7"} Dec 07 09:31:38 crc kubenswrapper[4838]: I1207 09:31:38.973148 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vps95" event={"ID":"b2ac318d-94b4-4ba4-8944-d597ce82db2e","Type":"ContainerStarted","Data":"5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea"} Dec 07 09:31:38 crc kubenswrapper[4838]: I1207 09:31:38.995266 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vps95" podStartSLOduration=2.620730211 podStartE2EDuration="5.995248273s" podCreationTimestamp="2025-12-07 09:31:33 +0000 UTC" firstStartedPulling="2025-12-07 09:31:34.937224914 +0000 UTC m=+1511.644543931" lastFinishedPulling="2025-12-07 09:31:38.311742976 +0000 UTC m=+1515.019061993" observedRunningTime="2025-12-07 09:31:38.991936229 +0000 UTC m=+1515.699255266" watchObservedRunningTime="2025-12-07 09:31:38.995248273 +0000 UTC m=+1515.702567290" Dec 07 09:31:44 crc kubenswrapper[4838]: I1207 09:31:44.102951 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:44 crc kubenswrapper[4838]: I1207 09:31:44.103443 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:44 crc kubenswrapper[4838]: I1207 09:31:44.154755 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:45 crc kubenswrapper[4838]: I1207 09:31:45.077213 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:45 crc kubenswrapper[4838]: I1207 09:31:45.123667 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vps95"] Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.049167 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vps95" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="registry-server" containerID="cri-o://5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea" gracePeriod=2 Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.504245 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.541063 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-catalog-content\") pod \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.541226 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhltd\" (UniqueName: \"kubernetes.io/projected/b2ac318d-94b4-4ba4-8944-d597ce82db2e-kube-api-access-lhltd\") pod \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.541439 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-utilities\") pod \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\" (UID: \"b2ac318d-94b4-4ba4-8944-d597ce82db2e\") " Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.542172 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-utilities" (OuterVolumeSpecName: "utilities") pod "b2ac318d-94b4-4ba4-8944-d597ce82db2e" (UID: "b2ac318d-94b4-4ba4-8944-d597ce82db2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.546384 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ac318d-94b4-4ba4-8944-d597ce82db2e-kube-api-access-lhltd" (OuterVolumeSpecName: "kube-api-access-lhltd") pod "b2ac318d-94b4-4ba4-8944-d597ce82db2e" (UID: "b2ac318d-94b4-4ba4-8944-d597ce82db2e"). InnerVolumeSpecName "kube-api-access-lhltd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.604635 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2ac318d-94b4-4ba4-8944-d597ce82db2e" (UID: "b2ac318d-94b4-4ba4-8944-d597ce82db2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.643402 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.643433 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ac318d-94b4-4ba4-8944-d597ce82db2e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:31:47 crc kubenswrapper[4838]: I1207 09:31:47.643446 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhltd\" (UniqueName: \"kubernetes.io/projected/b2ac318d-94b4-4ba4-8944-d597ce82db2e-kube-api-access-lhltd\") on node \"crc\" DevicePath \"\"" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.061226 4838 generic.go:334] "Generic (PLEG): container finished" podID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerID="5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea" exitCode=0 Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.061300 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vps95" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.061307 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vps95" event={"ID":"b2ac318d-94b4-4ba4-8944-d597ce82db2e","Type":"ContainerDied","Data":"5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea"} Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.061376 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vps95" event={"ID":"b2ac318d-94b4-4ba4-8944-d597ce82db2e","Type":"ContainerDied","Data":"2da47d3267cfb02bbeeb6148c6ee7c914e2e1165da79d00c8255ff61d8f14f9b"} Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.061396 4838 scope.go:117] "RemoveContainer" containerID="5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.093425 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vps95"] Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.108406 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vps95"] Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.116067 4838 scope.go:117] "RemoveContainer" containerID="82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.163175 4838 scope.go:117] "RemoveContainer" containerID="df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.200710 4838 scope.go:117] "RemoveContainer" containerID="5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea" Dec 07 09:31:48 crc kubenswrapper[4838]: E1207 09:31:48.201391 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea\": container with ID starting with 5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea not found: ID does not exist" containerID="5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.201423 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea"} err="failed to get container status \"5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea\": rpc error: code = NotFound desc = could not find container \"5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea\": container with ID starting with 5f630ec6036dee3f33424f433d4226e2f3d543653d2434f3130cd59cae37f6ea not found: ID does not exist" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.201443 4838 scope.go:117] "RemoveContainer" containerID="82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7" Dec 07 09:31:48 crc kubenswrapper[4838]: E1207 09:31:48.201873 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7\": container with ID starting with 82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7 not found: ID does not exist" containerID="82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.202000 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7"} err="failed to get container status \"82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7\": rpc error: code = NotFound desc = could not find container \"82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7\": container with ID starting with 82fbbfe9fdba041da386e3abcd3341da34063ac8deb2afc4447cbf1404c9d0b7 not found: ID does not exist" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.202110 4838 scope.go:117] "RemoveContainer" containerID="df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82" Dec 07 09:31:48 crc kubenswrapper[4838]: E1207 09:31:48.202509 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82\": container with ID starting with df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82 not found: ID does not exist" containerID="df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82" Dec 07 09:31:48 crc kubenswrapper[4838]: I1207 09:31:48.202538 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82"} err="failed to get container status \"df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82\": rpc error: code = NotFound desc = could not find container \"df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82\": container with ID starting with df65a6f1efd35b3b5b2c169269a8da3a81dc6357971201295925c4f98af3ea82 not found: ID does not exist" Dec 07 09:31:49 crc kubenswrapper[4838]: I1207 09:31:49.502228 4838 scope.go:117] "RemoveContainer" containerID="919e11cc574bf359d79f319331a94db2ff6666f563e4ce270263c73f88d7f826" Dec 07 09:31:49 crc kubenswrapper[4838]: I1207 09:31:49.534676 4838 scope.go:117] "RemoveContainer" containerID="33faa1be103a96fa5e0f624953e21f1c7d2519138bcd20ba9ad0c6610199ce06" Dec 07 09:31:49 crc kubenswrapper[4838]: I1207 09:31:49.582875 4838 scope.go:117] "RemoveContainer" containerID="116c573ecd2d9c117ba663a2ece45430c3599ffa21d65b00d3de6339f5f00f7c" Dec 07 09:31:49 crc kubenswrapper[4838]: I1207 09:31:49.627236 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" path="/var/lib/kubelet/pods/b2ac318d-94b4-4ba4-8944-d597ce82db2e/volumes" Dec 07 09:31:54 crc kubenswrapper[4838]: I1207 09:31:54.493572 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:31:54 crc kubenswrapper[4838]: I1207 09:31:54.494234 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:31:54 crc kubenswrapper[4838]: I1207 09:31:54.494289 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:31:54 crc kubenswrapper[4838]: I1207 09:31:54.495181 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:31:54 crc kubenswrapper[4838]: I1207 09:31:54.495250 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" gracePeriod=600 Dec 07 09:31:54 crc kubenswrapper[4838]: E1207 09:31:54.621795 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:31:55 crc kubenswrapper[4838]: I1207 09:31:55.119657 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" exitCode=0 Dec 07 09:31:55 crc kubenswrapper[4838]: I1207 09:31:55.119735 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3"} Dec 07 09:31:55 crc kubenswrapper[4838]: I1207 09:31:55.120074 4838 scope.go:117] "RemoveContainer" containerID="1c156b05b4ce4b0fa0902c6340a3fd6771fc44b04d8150317eefeaccfa29bca9" Dec 07 09:31:55 crc kubenswrapper[4838]: I1207 09:31:55.120751 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:31:55 crc kubenswrapper[4838]: E1207 09:31:55.121169 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.500697 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-84vcd"] Dec 07 09:31:58 crc kubenswrapper[4838]: E1207 09:31:58.501595 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="extract-content" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.501606 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="extract-content" Dec 07 09:31:58 crc kubenswrapper[4838]: E1207 09:31:58.501621 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="registry-server" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.501629 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="registry-server" Dec 07 09:31:58 crc kubenswrapper[4838]: E1207 09:31:58.501652 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="extract-utilities" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.501658 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="extract-utilities" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.501849 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ac318d-94b4-4ba4-8944-d597ce82db2e" containerName="registry-server" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.503103 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.516940 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84vcd"] Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.540028 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-catalog-content\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.540301 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-utilities\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.540554 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb8s5\" (UniqueName: \"kubernetes.io/projected/5eea708d-85d8-46f9-ba09-e6f44d001d4f-kube-api-access-wb8s5\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.642706 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb8s5\" (UniqueName: \"kubernetes.io/projected/5eea708d-85d8-46f9-ba09-e6f44d001d4f-kube-api-access-wb8s5\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.643019 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-catalog-content\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.643133 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-utilities\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.643563 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-catalog-content\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.643717 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-utilities\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.662267 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb8s5\" (UniqueName: \"kubernetes.io/projected/5eea708d-85d8-46f9-ba09-e6f44d001d4f-kube-api-access-wb8s5\") pod \"community-operators-84vcd\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:58 crc kubenswrapper[4838]: I1207 09:31:58.824842 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:31:59 crc kubenswrapper[4838]: I1207 09:31:59.321792 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84vcd"] Dec 07 09:32:00 crc kubenswrapper[4838]: I1207 09:32:00.188854 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerID="2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283" exitCode=0 Dec 07 09:32:00 crc kubenswrapper[4838]: I1207 09:32:00.189020 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84vcd" event={"ID":"5eea708d-85d8-46f9-ba09-e6f44d001d4f","Type":"ContainerDied","Data":"2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283"} Dec 07 09:32:00 crc kubenswrapper[4838]: I1207 09:32:00.189421 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84vcd" event={"ID":"5eea708d-85d8-46f9-ba09-e6f44d001d4f","Type":"ContainerStarted","Data":"e16164a306ddeaaec52d8b851ddc22bdda28fe5f75a75afbc79dcff86dd003fa"} Dec 07 09:32:01 crc kubenswrapper[4838]: I1207 09:32:01.199739 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84vcd" event={"ID":"5eea708d-85d8-46f9-ba09-e6f44d001d4f","Type":"ContainerStarted","Data":"ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d"} Dec 07 09:32:03 crc kubenswrapper[4838]: I1207 09:32:03.223163 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerID="ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d" exitCode=0 Dec 07 09:32:03 crc kubenswrapper[4838]: I1207 09:32:03.223299 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84vcd" event={"ID":"5eea708d-85d8-46f9-ba09-e6f44d001d4f","Type":"ContainerDied","Data":"ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d"} Dec 07 09:32:04 crc kubenswrapper[4838]: I1207 09:32:04.235695 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84vcd" event={"ID":"5eea708d-85d8-46f9-ba09-e6f44d001d4f","Type":"ContainerStarted","Data":"a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2"} Dec 07 09:32:04 crc kubenswrapper[4838]: I1207 09:32:04.272351 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-84vcd" podStartSLOduration=2.8430635349999998 podStartE2EDuration="6.272334594s" podCreationTimestamp="2025-12-07 09:31:58 +0000 UTC" firstStartedPulling="2025-12-07 09:32:00.191997581 +0000 UTC m=+1536.899316608" lastFinishedPulling="2025-12-07 09:32:03.62126865 +0000 UTC m=+1540.328587667" observedRunningTime="2025-12-07 09:32:04.257687227 +0000 UTC m=+1540.965006254" watchObservedRunningTime="2025-12-07 09:32:04.272334594 +0000 UTC m=+1540.979653611" Dec 07 09:32:06 crc kubenswrapper[4838]: I1207 09:32:06.614529 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:32:06 crc kubenswrapper[4838]: E1207 09:32:06.615045 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:32:08 crc kubenswrapper[4838]: I1207 09:32:08.825409 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:32:08 crc kubenswrapper[4838]: I1207 09:32:08.825755 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:32:08 crc kubenswrapper[4838]: I1207 09:32:08.883802 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:32:09 crc kubenswrapper[4838]: I1207 09:32:09.337367 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:32:09 crc kubenswrapper[4838]: I1207 09:32:09.387782 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84vcd"] Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.303861 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-84vcd" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="registry-server" containerID="cri-o://a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2" gracePeriod=2 Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.749630 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.801662 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb8s5\" (UniqueName: \"kubernetes.io/projected/5eea708d-85d8-46f9-ba09-e6f44d001d4f-kube-api-access-wb8s5\") pod \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.802349 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-catalog-content\") pod \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.802780 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-utilities\") pod \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\" (UID: \"5eea708d-85d8-46f9-ba09-e6f44d001d4f\") " Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.804314 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-utilities" (OuterVolumeSpecName: "utilities") pod "5eea708d-85d8-46f9-ba09-e6f44d001d4f" (UID: "5eea708d-85d8-46f9-ba09-e6f44d001d4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.805255 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.807864 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eea708d-85d8-46f9-ba09-e6f44d001d4f-kube-api-access-wb8s5" (OuterVolumeSpecName: "kube-api-access-wb8s5") pod "5eea708d-85d8-46f9-ba09-e6f44d001d4f" (UID: "5eea708d-85d8-46f9-ba09-e6f44d001d4f"). InnerVolumeSpecName "kube-api-access-wb8s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.851158 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5eea708d-85d8-46f9-ba09-e6f44d001d4f" (UID: "5eea708d-85d8-46f9-ba09-e6f44d001d4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.907111 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb8s5\" (UniqueName: \"kubernetes.io/projected/5eea708d-85d8-46f9-ba09-e6f44d001d4f-kube-api-access-wb8s5\") on node \"crc\" DevicePath \"\"" Dec 07 09:32:11 crc kubenswrapper[4838]: I1207 09:32:11.907148 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5eea708d-85d8-46f9-ba09-e6f44d001d4f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.321045 4838 generic.go:334] "Generic (PLEG): container finished" podID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerID="a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2" exitCode=0 Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.321107 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84vcd" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.321116 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84vcd" event={"ID":"5eea708d-85d8-46f9-ba09-e6f44d001d4f","Type":"ContainerDied","Data":"a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2"} Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.321984 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84vcd" event={"ID":"5eea708d-85d8-46f9-ba09-e6f44d001d4f","Type":"ContainerDied","Data":"e16164a306ddeaaec52d8b851ddc22bdda28fe5f75a75afbc79dcff86dd003fa"} Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.322030 4838 scope.go:117] "RemoveContainer" containerID="a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.360510 4838 scope.go:117] "RemoveContainer" containerID="ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.371413 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84vcd"] Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.381397 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-84vcd"] Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.392517 4838 scope.go:117] "RemoveContainer" containerID="2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.421644 4838 scope.go:117] "RemoveContainer" containerID="a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2" Dec 07 09:32:12 crc kubenswrapper[4838]: E1207 09:32:12.422081 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2\": container with ID starting with a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2 not found: ID does not exist" containerID="a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.422113 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2"} err="failed to get container status \"a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2\": rpc error: code = NotFound desc = could not find container \"a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2\": container with ID starting with a7a1676a6cecadbaae1378e1ad98f1704023afd5a472c3f9ed84cdebb24e88f2 not found: ID does not exist" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.422135 4838 scope.go:117] "RemoveContainer" containerID="ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d" Dec 07 09:32:12 crc kubenswrapper[4838]: E1207 09:32:12.422333 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d\": container with ID starting with ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d not found: ID does not exist" containerID="ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.422356 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d"} err="failed to get container status \"ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d\": rpc error: code = NotFound desc = could not find container \"ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d\": container with ID starting with ed3749ea93c8a353d5cb799891046f848794cbf2ed06e77f4be279e26edd951d not found: ID does not exist" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.422373 4838 scope.go:117] "RemoveContainer" containerID="2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283" Dec 07 09:32:12 crc kubenswrapper[4838]: E1207 09:32:12.422560 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283\": container with ID starting with 2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283 not found: ID does not exist" containerID="2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283" Dec 07 09:32:12 crc kubenswrapper[4838]: I1207 09:32:12.422586 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283"} err="failed to get container status \"2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283\": rpc error: code = NotFound desc = could not find container \"2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283\": container with ID starting with 2f17ac11e834cb53cb9cece8aee0b2c19bac912034275be7927e4a48c9172283 not found: ID does not exist" Dec 07 09:32:13 crc kubenswrapper[4838]: I1207 09:32:13.626988 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" path="/var/lib/kubelet/pods/5eea708d-85d8-46f9-ba09-e6f44d001d4f/volumes" Dec 07 09:32:20 crc kubenswrapper[4838]: I1207 09:32:20.614742 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:32:20 crc kubenswrapper[4838]: E1207 09:32:20.615547 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:32:33 crc kubenswrapper[4838]: I1207 09:32:33.628293 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:32:33 crc kubenswrapper[4838]: E1207 09:32:33.629051 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:32:45 crc kubenswrapper[4838]: I1207 09:32:45.615505 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:32:45 crc kubenswrapper[4838]: E1207 09:32:45.616870 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:33:00 crc kubenswrapper[4838]: I1207 09:33:00.614876 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:33:00 crc kubenswrapper[4838]: E1207 09:33:00.615623 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:33:15 crc kubenswrapper[4838]: I1207 09:33:15.615056 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:33:15 crc kubenswrapper[4838]: E1207 09:33:15.616035 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:33:26 crc kubenswrapper[4838]: I1207 09:33:26.614613 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:33:26 crc kubenswrapper[4838]: E1207 09:33:26.615406 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:33:40 crc kubenswrapper[4838]: I1207 09:33:40.615406 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:33:40 crc kubenswrapper[4838]: E1207 09:33:40.616279 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:33:46 crc kubenswrapper[4838]: I1207 09:33:46.328661 4838 generic.go:334] "Generic (PLEG): container finished" podID="9e993146-3e36-443e-b88b-1d090aa17900" containerID="19d85c270699df25aee8183fa0de7b595a3af3a8a51966a213675c01c342beb3" exitCode=0 Dec 07 09:33:46 crc kubenswrapper[4838]: I1207 09:33:46.328745 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" event={"ID":"9e993146-3e36-443e-b88b-1d090aa17900","Type":"ContainerDied","Data":"19d85c270699df25aee8183fa0de7b595a3af3a8a51966a213675c01c342beb3"} Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.764627 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.932777 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-ssh-key\") pod \"9e993146-3e36-443e-b88b-1d090aa17900\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.932923 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx9q2\" (UniqueName: \"kubernetes.io/projected/9e993146-3e36-443e-b88b-1d090aa17900-kube-api-access-sx9q2\") pod \"9e993146-3e36-443e-b88b-1d090aa17900\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.933589 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-inventory\") pod \"9e993146-3e36-443e-b88b-1d090aa17900\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.933667 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-bootstrap-combined-ca-bundle\") pod \"9e993146-3e36-443e-b88b-1d090aa17900\" (UID: \"9e993146-3e36-443e-b88b-1d090aa17900\") " Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.940084 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e993146-3e36-443e-b88b-1d090aa17900-kube-api-access-sx9q2" (OuterVolumeSpecName: "kube-api-access-sx9q2") pod "9e993146-3e36-443e-b88b-1d090aa17900" (UID: "9e993146-3e36-443e-b88b-1d090aa17900"). InnerVolumeSpecName "kube-api-access-sx9q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.940109 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9e993146-3e36-443e-b88b-1d090aa17900" (UID: "9e993146-3e36-443e-b88b-1d090aa17900"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.960904 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9e993146-3e36-443e-b88b-1d090aa17900" (UID: "9e993146-3e36-443e-b88b-1d090aa17900"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:33:47 crc kubenswrapper[4838]: I1207 09:33:47.973592 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-inventory" (OuterVolumeSpecName: "inventory") pod "9e993146-3e36-443e-b88b-1d090aa17900" (UID: "9e993146-3e36-443e-b88b-1d090aa17900"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.036249 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.036273 4838 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.036283 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e993146-3e36-443e-b88b-1d090aa17900-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.036293 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx9q2\" (UniqueName: \"kubernetes.io/projected/9e993146-3e36-443e-b88b-1d090aa17900-kube-api-access-sx9q2\") on node \"crc\" DevicePath \"\"" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.387783 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" event={"ID":"9e993146-3e36-443e-b88b-1d090aa17900","Type":"ContainerDied","Data":"64bda83a9f9966223defb6b92ca8f2bcbce472e4684c7cbbe7d83c0f87580e02"} Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.387834 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64bda83a9f9966223defb6b92ca8f2bcbce472e4684c7cbbe7d83c0f87580e02" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.387888 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.463183 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99"] Dec 07 09:33:48 crc kubenswrapper[4838]: E1207 09:33:48.463694 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="extract-content" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.463717 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="extract-content" Dec 07 09:33:48 crc kubenswrapper[4838]: E1207 09:33:48.463730 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="extract-utilities" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.463739 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="extract-utilities" Dec 07 09:33:48 crc kubenswrapper[4838]: E1207 09:33:48.463754 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e993146-3e36-443e-b88b-1d090aa17900" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.463763 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e993146-3e36-443e-b88b-1d090aa17900" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 09:33:48 crc kubenswrapper[4838]: E1207 09:33:48.463773 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="registry-server" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.463780 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="registry-server" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.464019 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eea708d-85d8-46f9-ba09-e6f44d001d4f" containerName="registry-server" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.464043 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e993146-3e36-443e-b88b-1d090aa17900" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.464667 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.470351 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.470432 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.470377 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.470524 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.473994 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99"] Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.546453 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.546525 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8t6k\" (UniqueName: \"kubernetes.io/projected/2bc69631-6ca5-4497-b711-aa0a2d568bbf-kube-api-access-n8t6k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.546586 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.648472 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.648564 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8t6k\" (UniqueName: \"kubernetes.io/projected/2bc69631-6ca5-4497-b711-aa0a2d568bbf-kube-api-access-n8t6k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.648620 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.653376 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.654886 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.677151 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8t6k\" (UniqueName: \"kubernetes.io/projected/2bc69631-6ca5-4497-b711-aa0a2d568bbf-kube-api-access-n8t6k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-l6q99\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:48 crc kubenswrapper[4838]: I1207 09:33:48.789421 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:33:49 crc kubenswrapper[4838]: I1207 09:33:49.315134 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99"] Dec 07 09:33:49 crc kubenswrapper[4838]: I1207 09:33:49.333362 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:33:49 crc kubenswrapper[4838]: I1207 09:33:49.398246 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" event={"ID":"2bc69631-6ca5-4497-b711-aa0a2d568bbf","Type":"ContainerStarted","Data":"eab206a6376540b153727a78b7eeb43c159ed9a8a9a19604227129de84d53835"} Dec 07 09:33:50 crc kubenswrapper[4838]: I1207 09:33:50.410608 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" event={"ID":"2bc69631-6ca5-4497-b711-aa0a2d568bbf","Type":"ContainerStarted","Data":"b6073d5b04ed3abcb072c8855b1b69665acbf3072734d569e28ac30635a5d783"} Dec 07 09:33:50 crc kubenswrapper[4838]: I1207 09:33:50.437337 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" podStartSLOduration=1.87319456 podStartE2EDuration="2.437314946s" podCreationTimestamp="2025-12-07 09:33:48 +0000 UTC" firstStartedPulling="2025-12-07 09:33:49.333102989 +0000 UTC m=+1646.040422026" lastFinishedPulling="2025-12-07 09:33:49.897223385 +0000 UTC m=+1646.604542412" observedRunningTime="2025-12-07 09:33:50.434128654 +0000 UTC m=+1647.141447721" watchObservedRunningTime="2025-12-07 09:33:50.437314946 +0000 UTC m=+1647.144633973" Dec 07 09:33:53 crc kubenswrapper[4838]: I1207 09:33:53.622677 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:33:53 crc kubenswrapper[4838]: E1207 09:33:53.623442 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:34:06 crc kubenswrapper[4838]: I1207 09:34:06.614591 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:34:06 crc kubenswrapper[4838]: E1207 09:34:06.615440 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:34:21 crc kubenswrapper[4838]: I1207 09:34:21.615068 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:34:21 crc kubenswrapper[4838]: E1207 09:34:21.615772 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:34:34 crc kubenswrapper[4838]: I1207 09:34:34.614925 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:34:34 crc kubenswrapper[4838]: E1207 09:34:34.615642 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:34:47 crc kubenswrapper[4838]: I1207 09:34:47.049331 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5nft8"] Dec 07 09:34:47 crc kubenswrapper[4838]: I1207 09:34:47.058577 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5nft8"] Dec 07 09:34:47 crc kubenswrapper[4838]: I1207 09:34:47.626319 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6369855a-6de3-482c-8b9c-3e0263b7979f" path="/var/lib/kubelet/pods/6369855a-6de3-482c-8b9c-3e0263b7979f/volumes" Dec 07 09:34:48 crc kubenswrapper[4838]: I1207 09:34:48.614430 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:34:48 crc kubenswrapper[4838]: E1207 09:34:48.614683 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.039743 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g9b52"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.048190 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-21a4-account-create-update-p44qz"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.058353 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g9b52"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.067643 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5jhzd"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.076111 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5jhzd"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.083933 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-1db7-account-create-update-6m8pz"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.091770 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-ecf1-account-create-update-k27fv"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.100477 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-21a4-account-create-update-p44qz"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.109056 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-1db7-account-create-update-6m8pz"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.117570 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-ecf1-account-create-update-k27fv"] Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.628541 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05852af0-5b52-4356-97e5-2fb3e9800c65" path="/var/lib/kubelet/pods/05852af0-5b52-4356-97e5-2fb3e9800c65/volumes" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.629612 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="298292bd-8186-4646-ad35-8058c9963946" path="/var/lib/kubelet/pods/298292bd-8186-4646-ad35-8058c9963946/volumes" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.630521 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="305e31d7-29e9-44f4-9dec-2aa2fd20c46c" path="/var/lib/kubelet/pods/305e31d7-29e9-44f4-9dec-2aa2fd20c46c/volumes" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.631422 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6863df38-7011-4203-adb9-117671f13bf4" path="/var/lib/kubelet/pods/6863df38-7011-4203-adb9-117671f13bf4/volumes" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.633148 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee0e6d6c-351d-4b99-82d1-b2ad1042e44e" path="/var/lib/kubelet/pods/ee0e6d6c-351d-4b99-82d1-b2ad1042e44e/volumes" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.806004 4838 scope.go:117] "RemoveContainer" containerID="157f46c3097105f4976e547241583cf8932c11f8a1ff77a88bd7deeaf723e27b" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.832960 4838 scope.go:117] "RemoveContainer" containerID="8082d8d1d591b7d583d27c21776e7508886c399bef9fdfed7413bfddb8a5cf47" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.898606 4838 scope.go:117] "RemoveContainer" containerID="4dc72005e29e40e16a332f0c82e3ea5dec6046c64ef78695f7f8b1cda5882398" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.925956 4838 scope.go:117] "RemoveContainer" containerID="3a81939f19a7888bbc7b1d965225c8d61c2b5dacb2513fc3c2531cf4b8c4cb07" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.949504 4838 scope.go:117] "RemoveContainer" containerID="2500aefe6979699207447925bb0966c448336da033795d2d1e03a87a69900cfb" Dec 07 09:34:49 crc kubenswrapper[4838]: I1207 09:34:49.991538 4838 scope.go:117] "RemoveContainer" containerID="30b95a7a6ec279e0cd2c3400099c92390b94d7f99da3170a4d4a204409f5d4bb" Dec 07 09:34:50 crc kubenswrapper[4838]: I1207 09:34:50.030660 4838 scope.go:117] "RemoveContainer" containerID="3dfe2c1926797d8cb64d8a414bc43c6b7980f204a0e84e8ebeab75650335e06b" Dec 07 09:34:50 crc kubenswrapper[4838]: I1207 09:34:50.066253 4838 scope.go:117] "RemoveContainer" containerID="2ba8fddc8a33dfaed82c785f863267073eb9608fab8ab17f0c11f872b5c6e6d8" Dec 07 09:35:00 crc kubenswrapper[4838]: I1207 09:35:00.615268 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:35:00 crc kubenswrapper[4838]: E1207 09:35:00.616430 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:35:10 crc kubenswrapper[4838]: I1207 09:35:10.055107 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xql5b"] Dec 07 09:35:10 crc kubenswrapper[4838]: I1207 09:35:10.064844 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xql5b"] Dec 07 09:35:11 crc kubenswrapper[4838]: I1207 09:35:11.047203 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-f574-account-create-update-8q9qx"] Dec 07 09:35:11 crc kubenswrapper[4838]: I1207 09:35:11.061350 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-f574-account-create-update-8q9qx"] Dec 07 09:35:11 crc kubenswrapper[4838]: I1207 09:35:11.090959 4838 generic.go:334] "Generic (PLEG): container finished" podID="2bc69631-6ca5-4497-b711-aa0a2d568bbf" containerID="b6073d5b04ed3abcb072c8855b1b69665acbf3072734d569e28ac30635a5d783" exitCode=0 Dec 07 09:35:11 crc kubenswrapper[4838]: I1207 09:35:11.091011 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" event={"ID":"2bc69631-6ca5-4497-b711-aa0a2d568bbf","Type":"ContainerDied","Data":"b6073d5b04ed3abcb072c8855b1b69665acbf3072734d569e28ac30635a5d783"} Dec 07 09:35:11 crc kubenswrapper[4838]: I1207 09:35:11.624400 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b524180a-e6ae-4cb3-9717-99ad184944b8" path="/var/lib/kubelet/pods/b524180a-e6ae-4cb3-9717-99ad184944b8/volumes" Dec 07 09:35:11 crc kubenswrapper[4838]: I1207 09:35:11.624962 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2a91f97-d03b-4da9-b70c-e858e878470f" path="/var/lib/kubelet/pods/c2a91f97-d03b-4da9-b70c-e858e878470f/volumes" Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.038216 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-cxm29"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.055588 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-kq6lf"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.069562 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-kq6lf"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.077264 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-72ed-account-create-update-tx6p4"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.084517 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-72ed-account-create-update-tx6p4"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.092397 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-cxm29"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.101772 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-800d-account-create-update-5qbg4"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.111216 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-800d-account-create-update-5qbg4"] Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.490157 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.589138 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-ssh-key\") pod \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.589236 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory\") pod \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.589285 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8t6k\" (UniqueName: \"kubernetes.io/projected/2bc69631-6ca5-4497-b711-aa0a2d568bbf-kube-api-access-n8t6k\") pod \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.594679 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc69631-6ca5-4497-b711-aa0a2d568bbf-kube-api-access-n8t6k" (OuterVolumeSpecName: "kube-api-access-n8t6k") pod "2bc69631-6ca5-4497-b711-aa0a2d568bbf" (UID: "2bc69631-6ca5-4497-b711-aa0a2d568bbf"). InnerVolumeSpecName "kube-api-access-n8t6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.617506 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:35:12 crc kubenswrapper[4838]: E1207 09:35:12.617814 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:35:12 crc kubenswrapper[4838]: E1207 09:35:12.620226 4838 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory podName:2bc69631-6ca5-4497-b711-aa0a2d568bbf nodeName:}" failed. No retries permitted until 2025-12-07 09:35:13.120199891 +0000 UTC m=+1729.827518898 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory") pod "2bc69631-6ca5-4497-b711-aa0a2d568bbf" (UID: "2bc69631-6ca5-4497-b711-aa0a2d568bbf") : error deleting /var/lib/kubelet/pods/2bc69631-6ca5-4497-b711-aa0a2d568bbf/volume-subpaths: remove /var/lib/kubelet/pods/2bc69631-6ca5-4497-b711-aa0a2d568bbf/volume-subpaths: no such file or directory Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.623542 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2bc69631-6ca5-4497-b711-aa0a2d568bbf" (UID: "2bc69631-6ca5-4497-b711-aa0a2d568bbf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.691027 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:35:12 crc kubenswrapper[4838]: I1207 09:35:12.691062 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8t6k\" (UniqueName: \"kubernetes.io/projected/2bc69631-6ca5-4497-b711-aa0a2d568bbf-kube-api-access-n8t6k\") on node \"crc\" DevicePath \"\"" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.111807 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" event={"ID":"2bc69631-6ca5-4497-b711-aa0a2d568bbf","Type":"ContainerDied","Data":"eab206a6376540b153727a78b7eeb43c159ed9a8a9a19604227129de84d53835"} Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.111887 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eab206a6376540b153727a78b7eeb43c159ed9a8a9a19604227129de84d53835" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.111893 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.199898 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory\") pod \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\" (UID: \"2bc69631-6ca5-4497-b711-aa0a2d568bbf\") " Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.203870 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6"] Dec 07 09:35:13 crc kubenswrapper[4838]: E1207 09:35:13.204336 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc69631-6ca5-4497-b711-aa0a2d568bbf" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.204357 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc69631-6ca5-4497-b711-aa0a2d568bbf" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.204553 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bc69631-6ca5-4497-b711-aa0a2d568bbf" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.205016 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory" (OuterVolumeSpecName: "inventory") pod "2bc69631-6ca5-4497-b711-aa0a2d568bbf" (UID: "2bc69631-6ca5-4497-b711-aa0a2d568bbf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.205293 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.222771 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6"] Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.302060 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.302106 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt6bh\" (UniqueName: \"kubernetes.io/projected/1af361e7-0c07-4c36-95fc-e53fc866b814-kube-api-access-zt6bh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.302144 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.302187 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bc69631-6ca5-4497-b711-aa0a2d568bbf-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.403754 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt6bh\" (UniqueName: \"kubernetes.io/projected/1af361e7-0c07-4c36-95fc-e53fc866b814-kube-api-access-zt6bh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.403794 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.403855 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.418745 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.419167 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.424100 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt6bh\" (UniqueName: \"kubernetes.io/projected/1af361e7-0c07-4c36-95fc-e53fc866b814-kube-api-access-zt6bh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.596391 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.627137 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="344ac587-0d87-4812-9d65-18fc79eec28b" path="/var/lib/kubelet/pods/344ac587-0d87-4812-9d65-18fc79eec28b/volumes" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.627879 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a38f9117-4655-46af-8b54-01ca6b4e6553" path="/var/lib/kubelet/pods/a38f9117-4655-46af-8b54-01ca6b4e6553/volumes" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.628559 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f" path="/var/lib/kubelet/pods/c03a56ab-6a42-4d81-a0bf-6cffbbcfb11f/volumes" Dec 07 09:35:13 crc kubenswrapper[4838]: I1207 09:35:13.629311 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c" path="/var/lib/kubelet/pods/dcb1b9bd-baf9-44fa-b43a-a181ffcaf89c/volumes" Dec 07 09:35:14 crc kubenswrapper[4838]: I1207 09:35:14.199715 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6"] Dec 07 09:35:15 crc kubenswrapper[4838]: I1207 09:35:15.129963 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" event={"ID":"1af361e7-0c07-4c36-95fc-e53fc866b814","Type":"ContainerStarted","Data":"cf921b0369ddf99c99cede84e0acf6c60c2005f8725e2b2af97ba7f5d8465570"} Dec 07 09:35:15 crc kubenswrapper[4838]: I1207 09:35:15.130255 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" event={"ID":"1af361e7-0c07-4c36-95fc-e53fc866b814","Type":"ContainerStarted","Data":"b823b08efd2206f23ab594aa954d4e2450f40608a78379dd8967237d98627c0e"} Dec 07 09:35:20 crc kubenswrapper[4838]: I1207 09:35:20.023888 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" podStartSLOduration=6.539907982 podStartE2EDuration="7.023865471s" podCreationTimestamp="2025-12-07 09:35:13 +0000 UTC" firstStartedPulling="2025-12-07 09:35:14.195427826 +0000 UTC m=+1730.902746843" lastFinishedPulling="2025-12-07 09:35:14.679385295 +0000 UTC m=+1731.386704332" observedRunningTime="2025-12-07 09:35:15.149444713 +0000 UTC m=+1731.856763740" watchObservedRunningTime="2025-12-07 09:35:20.023865471 +0000 UTC m=+1736.731184498" Dec 07 09:35:20 crc kubenswrapper[4838]: I1207 09:35:20.037382 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-rnxlk"] Dec 07 09:35:20 crc kubenswrapper[4838]: I1207 09:35:20.050673 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-n2jzr"] Dec 07 09:35:20 crc kubenswrapper[4838]: I1207 09:35:20.058392 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-rnxlk"] Dec 07 09:35:20 crc kubenswrapper[4838]: I1207 09:35:20.068030 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-n2jzr"] Dec 07 09:35:21 crc kubenswrapper[4838]: I1207 09:35:21.185339 4838 generic.go:334] "Generic (PLEG): container finished" podID="1af361e7-0c07-4c36-95fc-e53fc866b814" containerID="cf921b0369ddf99c99cede84e0acf6c60c2005f8725e2b2af97ba7f5d8465570" exitCode=0 Dec 07 09:35:21 crc kubenswrapper[4838]: I1207 09:35:21.185391 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" event={"ID":"1af361e7-0c07-4c36-95fc-e53fc866b814","Type":"ContainerDied","Data":"cf921b0369ddf99c99cede84e0acf6c60c2005f8725e2b2af97ba7f5d8465570"} Dec 07 09:35:21 crc kubenswrapper[4838]: I1207 09:35:21.627765 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa78bff1-c04d-42de-a3f1-7019500bd194" path="/var/lib/kubelet/pods/aa78bff1-c04d-42de-a3f1-7019500bd194/volumes" Dec 07 09:35:21 crc kubenswrapper[4838]: I1207 09:35:21.628365 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb3152f7-2ea4-4de8-ba34-038fd733682b" path="/var/lib/kubelet/pods/eb3152f7-2ea4-4de8-ba34-038fd733682b/volumes" Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.673460 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.781182 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-inventory\") pod \"1af361e7-0c07-4c36-95fc-e53fc866b814\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.781255 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt6bh\" (UniqueName: \"kubernetes.io/projected/1af361e7-0c07-4c36-95fc-e53fc866b814-kube-api-access-zt6bh\") pod \"1af361e7-0c07-4c36-95fc-e53fc866b814\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.781341 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-ssh-key\") pod \"1af361e7-0c07-4c36-95fc-e53fc866b814\" (UID: \"1af361e7-0c07-4c36-95fc-e53fc866b814\") " Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.802050 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af361e7-0c07-4c36-95fc-e53fc866b814-kube-api-access-zt6bh" (OuterVolumeSpecName: "kube-api-access-zt6bh") pod "1af361e7-0c07-4c36-95fc-e53fc866b814" (UID: "1af361e7-0c07-4c36-95fc-e53fc866b814"). InnerVolumeSpecName "kube-api-access-zt6bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.840145 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1af361e7-0c07-4c36-95fc-e53fc866b814" (UID: "1af361e7-0c07-4c36-95fc-e53fc866b814"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.870975 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-inventory" (OuterVolumeSpecName: "inventory") pod "1af361e7-0c07-4c36-95fc-e53fc866b814" (UID: "1af361e7-0c07-4c36-95fc-e53fc866b814"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.884886 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.884920 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt6bh\" (UniqueName: \"kubernetes.io/projected/1af361e7-0c07-4c36-95fc-e53fc866b814-kube-api-access-zt6bh\") on node \"crc\" DevicePath \"\"" Dec 07 09:35:22 crc kubenswrapper[4838]: I1207 09:35:22.884934 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1af361e7-0c07-4c36-95fc-e53fc866b814-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.205039 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" event={"ID":"1af361e7-0c07-4c36-95fc-e53fc866b814","Type":"ContainerDied","Data":"b823b08efd2206f23ab594aa954d4e2450f40608a78379dd8967237d98627c0e"} Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.205074 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b823b08efd2206f23ab594aa954d4e2450f40608a78379dd8967237d98627c0e" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.205160 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.287918 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247"] Dec 07 09:35:23 crc kubenswrapper[4838]: E1207 09:35:23.288372 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af361e7-0c07-4c36-95fc-e53fc866b814" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.288395 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af361e7-0c07-4c36-95fc-e53fc866b814" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.288601 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af361e7-0c07-4c36-95fc-e53fc866b814" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.289351 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.293201 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.293484 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.293705 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.293974 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.330032 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247"] Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.394292 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ckjp\" (UniqueName: \"kubernetes.io/projected/d2d32b7f-3264-4308-878b-874ccfd18e54-kube-api-access-2ckjp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.394460 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.394622 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.496481 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ckjp\" (UniqueName: \"kubernetes.io/projected/d2d32b7f-3264-4308-878b-874ccfd18e54-kube-api-access-2ckjp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.496578 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.496714 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.503370 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.503369 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.517633 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ckjp\" (UniqueName: \"kubernetes.io/projected/d2d32b7f-3264-4308-878b-874ccfd18e54-kube-api-access-2ckjp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gz247\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.634839 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:35:23 crc kubenswrapper[4838]: I1207 09:35:23.643218 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:35:24 crc kubenswrapper[4838]: I1207 09:35:24.160693 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247"] Dec 07 09:35:24 crc kubenswrapper[4838]: I1207 09:35:24.213536 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" event={"ID":"d2d32b7f-3264-4308-878b-874ccfd18e54","Type":"ContainerStarted","Data":"1c0122de50ee8cf9b5936a21ba66edb618184c754f298c178f58c2c50056e25a"} Dec 07 09:35:24 crc kubenswrapper[4838]: I1207 09:35:24.572224 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:35:25 crc kubenswrapper[4838]: I1207 09:35:25.222773 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" event={"ID":"d2d32b7f-3264-4308-878b-874ccfd18e54","Type":"ContainerStarted","Data":"df7a2b3f98440574d1a9abf0b1655c31d3286fdeec229e3e5a4fa10be3ed191b"} Dec 07 09:35:25 crc kubenswrapper[4838]: I1207 09:35:25.239263 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" podStartSLOduration=1.8366007930000001 podStartE2EDuration="2.239247482s" podCreationTimestamp="2025-12-07 09:35:23 +0000 UTC" firstStartedPulling="2025-12-07 09:35:24.167466694 +0000 UTC m=+1740.874785731" lastFinishedPulling="2025-12-07 09:35:24.570113403 +0000 UTC m=+1741.277432420" observedRunningTime="2025-12-07 09:35:25.236299393 +0000 UTC m=+1741.943618410" watchObservedRunningTime="2025-12-07 09:35:25.239247482 +0000 UTC m=+1741.946566499" Dec 07 09:35:26 crc kubenswrapper[4838]: I1207 09:35:26.615227 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:35:26 crc kubenswrapper[4838]: E1207 09:35:26.615866 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:35:38 crc kubenswrapper[4838]: I1207 09:35:38.615042 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:35:38 crc kubenswrapper[4838]: E1207 09:35:38.615894 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.193520 4838 scope.go:117] "RemoveContainer" containerID="513f58981ccaa9541916bf7c609645fb39ce6baf276662d9ca7a98f31aac9bef" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.275187 4838 scope.go:117] "RemoveContainer" containerID="60b2ac53a6c707e42f86895b236f2b648d2910e9a5365f1f0e18338f94c669c2" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.325843 4838 scope.go:117] "RemoveContainer" containerID="d4704ab1072a5d251238bb238e97298c48190866c5e02967d1cbe4f202871dd0" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.345536 4838 scope.go:117] "RemoveContainer" containerID="77b7638595e61d5be645e33e577c052000890a26a36f4590d525ac5e338be56e" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.384300 4838 scope.go:117] "RemoveContainer" containerID="a1063ec138e1b08dc2194a29b7f6adf93140bb578db8fea968b0c3cb3aadce30" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.435463 4838 scope.go:117] "RemoveContainer" containerID="b3b189385765253cf03e5f3046a7f6e53e471a9f3ab49b53deef23b341ab2e4d" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.477026 4838 scope.go:117] "RemoveContainer" containerID="366f9728572c990684bf4b94aca2d5853c6b60bf195c3f16de285ed8133c3a41" Dec 07 09:35:50 crc kubenswrapper[4838]: I1207 09:35:50.512624 4838 scope.go:117] "RemoveContainer" containerID="f014559b089827c91f803ee38dd392f758e3143e9253a23d798a766e65f58e3f" Dec 07 09:35:52 crc kubenswrapper[4838]: I1207 09:35:52.054988 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-nw8pl"] Dec 07 09:35:52 crc kubenswrapper[4838]: I1207 09:35:52.067991 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-nw8pl"] Dec 07 09:35:52 crc kubenswrapper[4838]: I1207 09:35:52.615897 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:35:52 crc kubenswrapper[4838]: E1207 09:35:52.616458 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:35:53 crc kubenswrapper[4838]: I1207 09:35:53.625539 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bbd28af-439a-4841-b3b8-7c984bbd9ee7" path="/var/lib/kubelet/pods/6bbd28af-439a-4841-b3b8-7c984bbd9ee7/volumes" Dec 07 09:35:54 crc kubenswrapper[4838]: I1207 09:35:54.062710 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-979hq"] Dec 07 09:35:54 crc kubenswrapper[4838]: I1207 09:35:54.082464 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-979hq"] Dec 07 09:35:55 crc kubenswrapper[4838]: I1207 09:35:55.625185 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8026163f-b7f8-4033-9742-79da3dde65a7" path="/var/lib/kubelet/pods/8026163f-b7f8-4033-9742-79da3dde65a7/volumes" Dec 07 09:35:56 crc kubenswrapper[4838]: I1207 09:35:56.052375 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-szwjk"] Dec 07 09:35:56 crc kubenswrapper[4838]: I1207 09:35:56.065806 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-szwjk"] Dec 07 09:35:57 crc kubenswrapper[4838]: I1207 09:35:57.629543 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="305f0d8c-8742-4318-b0be-9592574cee2f" path="/var/lib/kubelet/pods/305f0d8c-8742-4318-b0be-9592574cee2f/volumes" Dec 07 09:36:06 crc kubenswrapper[4838]: I1207 09:36:06.614933 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:36:06 crc kubenswrapper[4838]: E1207 09:36:06.615709 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:36:08 crc kubenswrapper[4838]: I1207 09:36:08.590316 4838 generic.go:334] "Generic (PLEG): container finished" podID="d2d32b7f-3264-4308-878b-874ccfd18e54" containerID="df7a2b3f98440574d1a9abf0b1655c31d3286fdeec229e3e5a4fa10be3ed191b" exitCode=0 Dec 07 09:36:08 crc kubenswrapper[4838]: I1207 09:36:08.590419 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" event={"ID":"d2d32b7f-3264-4308-878b-874ccfd18e54","Type":"ContainerDied","Data":"df7a2b3f98440574d1a9abf0b1655c31d3286fdeec229e3e5a4fa10be3ed191b"} Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.029119 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.150370 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-inventory\") pod \"d2d32b7f-3264-4308-878b-874ccfd18e54\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.150451 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ckjp\" (UniqueName: \"kubernetes.io/projected/d2d32b7f-3264-4308-878b-874ccfd18e54-kube-api-access-2ckjp\") pod \"d2d32b7f-3264-4308-878b-874ccfd18e54\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.150630 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-ssh-key\") pod \"d2d32b7f-3264-4308-878b-874ccfd18e54\" (UID: \"d2d32b7f-3264-4308-878b-874ccfd18e54\") " Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.156686 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d32b7f-3264-4308-878b-874ccfd18e54-kube-api-access-2ckjp" (OuterVolumeSpecName: "kube-api-access-2ckjp") pod "d2d32b7f-3264-4308-878b-874ccfd18e54" (UID: "d2d32b7f-3264-4308-878b-874ccfd18e54"). InnerVolumeSpecName "kube-api-access-2ckjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.212986 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-inventory" (OuterVolumeSpecName: "inventory") pod "d2d32b7f-3264-4308-878b-874ccfd18e54" (UID: "d2d32b7f-3264-4308-878b-874ccfd18e54"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.227014 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2d32b7f-3264-4308-878b-874ccfd18e54" (UID: "d2d32b7f-3264-4308-878b-874ccfd18e54"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.257843 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.257890 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d32b7f-3264-4308-878b-874ccfd18e54-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.257900 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ckjp\" (UniqueName: \"kubernetes.io/projected/d2d32b7f-3264-4308-878b-874ccfd18e54-kube-api-access-2ckjp\") on node \"crc\" DevicePath \"\"" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.606778 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" event={"ID":"d2d32b7f-3264-4308-878b-874ccfd18e54","Type":"ContainerDied","Data":"1c0122de50ee8cf9b5936a21ba66edb618184c754f298c178f58c2c50056e25a"} Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.606844 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.606849 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c0122de50ee8cf9b5936a21ba66edb618184c754f298c178f58c2c50056e25a" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.700229 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w"] Dec 07 09:36:10 crc kubenswrapper[4838]: E1207 09:36:10.700691 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d32b7f-3264-4308-878b-874ccfd18e54" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.700711 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d32b7f-3264-4308-878b-874ccfd18e54" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.700962 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d32b7f-3264-4308-878b-874ccfd18e54" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.701760 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.703932 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.704471 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.704536 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.709895 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.710312 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w"] Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.867621 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.867958 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.867980 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxsm\" (UniqueName: \"kubernetes.io/projected/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-kube-api-access-cpxsm\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.969270 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.969356 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.969385 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxsm\" (UniqueName: \"kubernetes.io/projected/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-kube-api-access-cpxsm\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.978547 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.978567 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:10 crc kubenswrapper[4838]: I1207 09:36:10.984554 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxsm\" (UniqueName: \"kubernetes.io/projected/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-kube-api-access-cpxsm\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.020771 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.058735 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-zhq8j"] Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.072986 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-zhq8j"] Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.080115 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-p9tch"] Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.086636 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-p9tch"] Dec 07 09:36:11 crc kubenswrapper[4838]: W1207 09:36:11.563632 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2c6eb84_9eb9_4101_9de7_dc74da6f4757.slice/crio-a09a69ff6d8f01750ca09be94fb26fb7e4619f56bf32b8796290475d01b114b1 WatchSource:0}: Error finding container a09a69ff6d8f01750ca09be94fb26fb7e4619f56bf32b8796290475d01b114b1: Status 404 returned error can't find the container with id a09a69ff6d8f01750ca09be94fb26fb7e4619f56bf32b8796290475d01b114b1 Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.572311 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w"] Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.623147 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081018e6-b090-43c9-a968-bfbf5edd56ff" path="/var/lib/kubelet/pods/081018e6-b090-43c9-a968-bfbf5edd56ff/volumes" Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.623879 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bee89ed4-9590-461e-9416-a4f0e9cad8b6" path="/var/lib/kubelet/pods/bee89ed4-9590-461e-9416-a4f0e9cad8b6/volumes" Dec 07 09:36:11 crc kubenswrapper[4838]: I1207 09:36:11.624443 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" event={"ID":"c2c6eb84-9eb9-4101-9de7-dc74da6f4757","Type":"ContainerStarted","Data":"a09a69ff6d8f01750ca09be94fb26fb7e4619f56bf32b8796290475d01b114b1"} Dec 07 09:36:12 crc kubenswrapper[4838]: I1207 09:36:12.626053 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" event={"ID":"c2c6eb84-9eb9-4101-9de7-dc74da6f4757","Type":"ContainerStarted","Data":"5ed677d43f8bccdeaafc520051c7efccfc880e4a12cd32f48bf2014cd801e02a"} Dec 07 09:36:12 crc kubenswrapper[4838]: I1207 09:36:12.647406 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" podStartSLOduration=2.098386151 podStartE2EDuration="2.647387253s" podCreationTimestamp="2025-12-07 09:36:10 +0000 UTC" firstStartedPulling="2025-12-07 09:36:11.566569611 +0000 UTC m=+1788.273888628" lastFinishedPulling="2025-12-07 09:36:12.115570713 +0000 UTC m=+1788.822889730" observedRunningTime="2025-12-07 09:36:12.646437367 +0000 UTC m=+1789.353756424" watchObservedRunningTime="2025-12-07 09:36:12.647387253 +0000 UTC m=+1789.354706270" Dec 07 09:36:17 crc kubenswrapper[4838]: I1207 09:36:17.615349 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:36:17 crc kubenswrapper[4838]: E1207 09:36:17.616306 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:36:17 crc kubenswrapper[4838]: I1207 09:36:17.670585 4838 generic.go:334] "Generic (PLEG): container finished" podID="c2c6eb84-9eb9-4101-9de7-dc74da6f4757" containerID="5ed677d43f8bccdeaafc520051c7efccfc880e4a12cd32f48bf2014cd801e02a" exitCode=0 Dec 07 09:36:17 crc kubenswrapper[4838]: I1207 09:36:17.670671 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" event={"ID":"c2c6eb84-9eb9-4101-9de7-dc74da6f4757","Type":"ContainerDied","Data":"5ed677d43f8bccdeaafc520051c7efccfc880e4a12cd32f48bf2014cd801e02a"} Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.083066 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.226941 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-ssh-key\") pod \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.227125 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-inventory\") pod \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.227173 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpxsm\" (UniqueName: \"kubernetes.io/projected/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-kube-api-access-cpxsm\") pod \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\" (UID: \"c2c6eb84-9eb9-4101-9de7-dc74da6f4757\") " Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.243072 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-kube-api-access-cpxsm" (OuterVolumeSpecName: "kube-api-access-cpxsm") pod "c2c6eb84-9eb9-4101-9de7-dc74da6f4757" (UID: "c2c6eb84-9eb9-4101-9de7-dc74da6f4757"). InnerVolumeSpecName "kube-api-access-cpxsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.275036 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-inventory" (OuterVolumeSpecName: "inventory") pod "c2c6eb84-9eb9-4101-9de7-dc74da6f4757" (UID: "c2c6eb84-9eb9-4101-9de7-dc74da6f4757"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.277214 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2c6eb84-9eb9-4101-9de7-dc74da6f4757" (UID: "c2c6eb84-9eb9-4101-9de7-dc74da6f4757"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.329167 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.329673 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.329763 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpxsm\" (UniqueName: \"kubernetes.io/projected/c2c6eb84-9eb9-4101-9de7-dc74da6f4757-kube-api-access-cpxsm\") on node \"crc\" DevicePath \"\"" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.689924 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.689853 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w" event={"ID":"c2c6eb84-9eb9-4101-9de7-dc74da6f4757","Type":"ContainerDied","Data":"a09a69ff6d8f01750ca09be94fb26fb7e4619f56bf32b8796290475d01b114b1"} Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.690695 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a09a69ff6d8f01750ca09be94fb26fb7e4619f56bf32b8796290475d01b114b1" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.780280 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx"] Dec 07 09:36:19 crc kubenswrapper[4838]: E1207 09:36:19.780876 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c6eb84-9eb9-4101-9de7-dc74da6f4757" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.780985 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c6eb84-9eb9-4101-9de7-dc74da6f4757" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.781282 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c6eb84-9eb9-4101-9de7-dc74da6f4757" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.782118 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.786232 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.786434 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.786647 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.786896 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.830734 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx"] Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.938751 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjnfb\" (UniqueName: \"kubernetes.io/projected/05413639-6dc7-443c-9300-975b60426883-kube-api-access-zjnfb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.939143 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:19 crc kubenswrapper[4838]: I1207 09:36:19.939343 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.041316 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.041546 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.042532 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjnfb\" (UniqueName: \"kubernetes.io/projected/05413639-6dc7-443c-9300-975b60426883-kube-api-access-zjnfb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.046269 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.052484 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.066599 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjnfb\" (UniqueName: \"kubernetes.io/projected/05413639-6dc7-443c-9300-975b60426883-kube-api-access-zjnfb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rppbx\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.141052 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:36:20 crc kubenswrapper[4838]: I1207 09:36:20.726587 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx"] Dec 07 09:36:21 crc kubenswrapper[4838]: I1207 09:36:21.708204 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" event={"ID":"05413639-6dc7-443c-9300-975b60426883","Type":"ContainerStarted","Data":"d1ce0601d6eaf316219ec31f5880f8ee80f783397491ed2653377f89677e5455"} Dec 07 09:36:21 crc kubenswrapper[4838]: I1207 09:36:21.708769 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" event={"ID":"05413639-6dc7-443c-9300-975b60426883","Type":"ContainerStarted","Data":"580d7626ee9f22f258826a6144eb2af80fbbc70367228403374cb81d33ef109d"} Dec 07 09:36:21 crc kubenswrapper[4838]: I1207 09:36:21.737576 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" podStartSLOduration=2.31021463 podStartE2EDuration="2.737557757s" podCreationTimestamp="2025-12-07 09:36:19 +0000 UTC" firstStartedPulling="2025-12-07 09:36:20.725702626 +0000 UTC m=+1797.433021643" lastFinishedPulling="2025-12-07 09:36:21.153045733 +0000 UTC m=+1797.860364770" observedRunningTime="2025-12-07 09:36:21.728016331 +0000 UTC m=+1798.435335348" watchObservedRunningTime="2025-12-07 09:36:21.737557757 +0000 UTC m=+1798.444876774" Dec 07 09:36:29 crc kubenswrapper[4838]: I1207 09:36:29.615877 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:36:29 crc kubenswrapper[4838]: E1207 09:36:29.623376 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:36:44 crc kubenswrapper[4838]: I1207 09:36:44.614315 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:36:44 crc kubenswrapper[4838]: E1207 09:36:44.615018 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:36:50 crc kubenswrapper[4838]: I1207 09:36:50.674025 4838 scope.go:117] "RemoveContainer" containerID="3828d64e8e66d54771c3226510dcf2721798e4507cf9459bc6924f8e491b76c2" Dec 07 09:36:50 crc kubenswrapper[4838]: I1207 09:36:50.738008 4838 scope.go:117] "RemoveContainer" containerID="251722eb59ee3717debbded5a41882aeda6424519869258a500aab70e6b0786e" Dec 07 09:36:50 crc kubenswrapper[4838]: I1207 09:36:50.768238 4838 scope.go:117] "RemoveContainer" containerID="41ba3ec9a51af44db44c6f8f91c9297a60544b71463e959c481ce09f78437310" Dec 07 09:36:50 crc kubenswrapper[4838]: I1207 09:36:50.830052 4838 scope.go:117] "RemoveContainer" containerID="8988d648c2a4f2568a8b55a738c09d6afe8e990870c687e7ca003c633938a21e" Dec 07 09:36:50 crc kubenswrapper[4838]: I1207 09:36:50.865207 4838 scope.go:117] "RemoveContainer" containerID="e64cad8549d04bedf8f36ef7d6db865d6c6840328c438c78ffcd458105234b9b" Dec 07 09:36:58 crc kubenswrapper[4838]: I1207 09:36:58.614556 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.028246 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"81353be09f7749f5b2e22eca0a163e7ec79346eb5cbc71ad52479317f9293c04"} Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.052999 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-ztr5c"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.061237 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-c71d-account-create-update-dtbgj"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.069760 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-98sxs"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.076662 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-975b7"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.083935 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1f91-account-create-update-gqmgj"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.092159 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ef37-account-create-update-djqnb"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.112573 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-98sxs"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.119648 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-975b7"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.134091 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-ztr5c"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.141150 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-c71d-account-create-update-dtbgj"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.155077 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1f91-account-create-update-gqmgj"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.161538 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ef37-account-create-update-djqnb"] Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.634411 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c25b096-cc08-4a86-bc3c-b5c8601b3f71" path="/var/lib/kubelet/pods/6c25b096-cc08-4a86-bc3c-b5c8601b3f71/volumes" Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.635613 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="779f6638-c56b-462b-a574-21d3a2e08c28" path="/var/lib/kubelet/pods/779f6638-c56b-462b-a574-21d3a2e08c28/volumes" Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.636342 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe4a29d-3f3d-4cc3-884f-00f111d7d651" path="/var/lib/kubelet/pods/7fe4a29d-3f3d-4cc3-884f-00f111d7d651/volumes" Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.637125 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e334fc-b52e-4b90-bcd6-736bdc1fcd3a" path="/var/lib/kubelet/pods/90e334fc-b52e-4b90-bcd6-736bdc1fcd3a/volumes" Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.638435 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e90708-7b16-4279-bb16-423c616b152c" path="/var/lib/kubelet/pods/99e90708-7b16-4279-bb16-423c616b152c/volumes" Dec 07 09:36:59 crc kubenswrapper[4838]: I1207 09:36:59.639250 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd80aba1-e836-4c0f-8eba-2999e284db44" path="/var/lib/kubelet/pods/dd80aba1-e836-4c0f-8eba-2999e284db44/volumes" Dec 07 09:37:09 crc kubenswrapper[4838]: I1207 09:37:09.947100 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-htpjr"] Dec 07 09:37:09 crc kubenswrapper[4838]: I1207 09:37:09.949617 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:09 crc kubenswrapper[4838]: I1207 09:37:09.956233 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-catalog-content\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:09 crc kubenswrapper[4838]: I1207 09:37:09.956284 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-utilities\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:09 crc kubenswrapper[4838]: I1207 09:37:09.956416 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvgl\" (UniqueName: \"kubernetes.io/projected/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-kube-api-access-8rvgl\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:09 crc kubenswrapper[4838]: I1207 09:37:09.969981 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-htpjr"] Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.057755 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-catalog-content\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.057802 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-utilities\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.057884 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvgl\" (UniqueName: \"kubernetes.io/projected/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-kube-api-access-8rvgl\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.058254 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-catalog-content\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.058337 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-utilities\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.077430 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvgl\" (UniqueName: \"kubernetes.io/projected/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-kube-api-access-8rvgl\") pod \"redhat-marketplace-htpjr\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.276031 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:10 crc kubenswrapper[4838]: I1207 09:37:10.728480 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-htpjr"] Dec 07 09:37:11 crc kubenswrapper[4838]: I1207 09:37:11.152202 4838 generic.go:334] "Generic (PLEG): container finished" podID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerID="87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93" exitCode=0 Dec 07 09:37:11 crc kubenswrapper[4838]: I1207 09:37:11.152243 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htpjr" event={"ID":"8ea19715-76c3-4ed3-8e41-ab30f222a3ac","Type":"ContainerDied","Data":"87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93"} Dec 07 09:37:11 crc kubenswrapper[4838]: I1207 09:37:11.152267 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htpjr" event={"ID":"8ea19715-76c3-4ed3-8e41-ab30f222a3ac","Type":"ContainerStarted","Data":"c14d343634feec6f2e154bbf113b0a8da5ad14fc596b36985c7a022cf3435e45"} Dec 07 09:37:12 crc kubenswrapper[4838]: I1207 09:37:12.172145 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htpjr" event={"ID":"8ea19715-76c3-4ed3-8e41-ab30f222a3ac","Type":"ContainerStarted","Data":"e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044"} Dec 07 09:37:13 crc kubenswrapper[4838]: I1207 09:37:13.181743 4838 generic.go:334] "Generic (PLEG): container finished" podID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerID="e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044" exitCode=0 Dec 07 09:37:13 crc kubenswrapper[4838]: I1207 09:37:13.182986 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htpjr" event={"ID":"8ea19715-76c3-4ed3-8e41-ab30f222a3ac","Type":"ContainerDied","Data":"e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044"} Dec 07 09:37:14 crc kubenswrapper[4838]: I1207 09:37:14.191948 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htpjr" event={"ID":"8ea19715-76c3-4ed3-8e41-ab30f222a3ac","Type":"ContainerStarted","Data":"ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff"} Dec 07 09:37:14 crc kubenswrapper[4838]: I1207 09:37:14.213403 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-htpjr" podStartSLOduration=2.739399885 podStartE2EDuration="5.213382428s" podCreationTimestamp="2025-12-07 09:37:09 +0000 UTC" firstStartedPulling="2025-12-07 09:37:11.155120733 +0000 UTC m=+1847.862439750" lastFinishedPulling="2025-12-07 09:37:13.629103276 +0000 UTC m=+1850.336422293" observedRunningTime="2025-12-07 09:37:14.207665713 +0000 UTC m=+1850.914984720" watchObservedRunningTime="2025-12-07 09:37:14.213382428 +0000 UTC m=+1850.920701445" Dec 07 09:37:20 crc kubenswrapper[4838]: I1207 09:37:20.276440 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:20 crc kubenswrapper[4838]: I1207 09:37:20.276989 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:20 crc kubenswrapper[4838]: I1207 09:37:20.329586 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:21 crc kubenswrapper[4838]: I1207 09:37:21.315009 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:21 crc kubenswrapper[4838]: I1207 09:37:21.375693 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-htpjr"] Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.062968 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8285b"] Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.071015 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8285b"] Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.285460 4838 generic.go:334] "Generic (PLEG): container finished" podID="05413639-6dc7-443c-9300-975b60426883" containerID="d1ce0601d6eaf316219ec31f5880f8ee80f783397491ed2653377f89677e5455" exitCode=0 Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.285558 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" event={"ID":"05413639-6dc7-443c-9300-975b60426883","Type":"ContainerDied","Data":"d1ce0601d6eaf316219ec31f5880f8ee80f783397491ed2653377f89677e5455"} Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.286153 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-htpjr" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="registry-server" containerID="cri-o://ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff" gracePeriod=2 Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.657985 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84318a4-61ad-4c08-928e-d06c3a99cc0b" path="/var/lib/kubelet/pods/d84318a4-61ad-4c08-928e-d06c3a99cc0b/volumes" Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.754792 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.840198 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-utilities\") pod \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.840292 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-catalog-content\") pod \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.840370 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rvgl\" (UniqueName: \"kubernetes.io/projected/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-kube-api-access-8rvgl\") pod \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\" (UID: \"8ea19715-76c3-4ed3-8e41-ab30f222a3ac\") " Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.841518 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-utilities" (OuterVolumeSpecName: "utilities") pod "8ea19715-76c3-4ed3-8e41-ab30f222a3ac" (UID: "8ea19715-76c3-4ed3-8e41-ab30f222a3ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.859057 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ea19715-76c3-4ed3-8e41-ab30f222a3ac" (UID: "8ea19715-76c3-4ed3-8e41-ab30f222a3ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.861792 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-kube-api-access-8rvgl" (OuterVolumeSpecName: "kube-api-access-8rvgl") pod "8ea19715-76c3-4ed3-8e41-ab30f222a3ac" (UID: "8ea19715-76c3-4ed3-8e41-ab30f222a3ac"). InnerVolumeSpecName "kube-api-access-8rvgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.942161 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rvgl\" (UniqueName: \"kubernetes.io/projected/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-kube-api-access-8rvgl\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.942198 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:23 crc kubenswrapper[4838]: I1207 09:37:23.942208 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea19715-76c3-4ed3-8e41-ab30f222a3ac-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.296907 4838 generic.go:334] "Generic (PLEG): container finished" podID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerID="ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff" exitCode=0 Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.296977 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htpjr" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.296973 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htpjr" event={"ID":"8ea19715-76c3-4ed3-8e41-ab30f222a3ac","Type":"ContainerDied","Data":"ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff"} Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.298120 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htpjr" event={"ID":"8ea19715-76c3-4ed3-8e41-ab30f222a3ac","Type":"ContainerDied","Data":"c14d343634feec6f2e154bbf113b0a8da5ad14fc596b36985c7a022cf3435e45"} Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.298182 4838 scope.go:117] "RemoveContainer" containerID="ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.327043 4838 scope.go:117] "RemoveContainer" containerID="e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.355933 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-htpjr"] Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.372529 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-htpjr"] Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.373476 4838 scope.go:117] "RemoveContainer" containerID="87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.419380 4838 scope.go:117] "RemoveContainer" containerID="ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff" Dec 07 09:37:24 crc kubenswrapper[4838]: E1207 09:37:24.420920 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff\": container with ID starting with ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff not found: ID does not exist" containerID="ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.420969 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff"} err="failed to get container status \"ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff\": rpc error: code = NotFound desc = could not find container \"ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff\": container with ID starting with ff6ee6502fc75c50adaffc131f93d7bb339cf7d1ae0a16877577f5434ca266ff not found: ID does not exist" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.421004 4838 scope.go:117] "RemoveContainer" containerID="e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044" Dec 07 09:37:24 crc kubenswrapper[4838]: E1207 09:37:24.421358 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044\": container with ID starting with e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044 not found: ID does not exist" containerID="e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.421395 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044"} err="failed to get container status \"e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044\": rpc error: code = NotFound desc = could not find container \"e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044\": container with ID starting with e6869e0494bf233b9e7c8a5172140381d6fa1c9a51d61c72338e568688718044 not found: ID does not exist" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.421420 4838 scope.go:117] "RemoveContainer" containerID="87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93" Dec 07 09:37:24 crc kubenswrapper[4838]: E1207 09:37:24.421677 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93\": container with ID starting with 87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93 not found: ID does not exist" containerID="87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.421715 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93"} err="failed to get container status \"87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93\": rpc error: code = NotFound desc = could not find container \"87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93\": container with ID starting with 87eb7fad56ef71a9ac3ed7ae3ecd77f38ce4d0453e85c2634892955805ca2b93 not found: ID does not exist" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.718454 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.755249 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-ssh-key\") pod \"05413639-6dc7-443c-9300-975b60426883\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.755326 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjnfb\" (UniqueName: \"kubernetes.io/projected/05413639-6dc7-443c-9300-975b60426883-kube-api-access-zjnfb\") pod \"05413639-6dc7-443c-9300-975b60426883\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.755391 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-inventory\") pod \"05413639-6dc7-443c-9300-975b60426883\" (UID: \"05413639-6dc7-443c-9300-975b60426883\") " Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.769962 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05413639-6dc7-443c-9300-975b60426883-kube-api-access-zjnfb" (OuterVolumeSpecName: "kube-api-access-zjnfb") pod "05413639-6dc7-443c-9300-975b60426883" (UID: "05413639-6dc7-443c-9300-975b60426883"). InnerVolumeSpecName "kube-api-access-zjnfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.783161 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "05413639-6dc7-443c-9300-975b60426883" (UID: "05413639-6dc7-443c-9300-975b60426883"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.783676 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-inventory" (OuterVolumeSpecName: "inventory") pod "05413639-6dc7-443c-9300-975b60426883" (UID: "05413639-6dc7-443c-9300-975b60426883"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.856760 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.856787 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05413639-6dc7-443c-9300-975b60426883-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:24 crc kubenswrapper[4838]: I1207 09:37:24.856798 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjnfb\" (UniqueName: \"kubernetes.io/projected/05413639-6dc7-443c-9300-975b60426883-kube-api-access-zjnfb\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.310525 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.312750 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx" event={"ID":"05413639-6dc7-443c-9300-975b60426883","Type":"ContainerDied","Data":"580d7626ee9f22f258826a6144eb2af80fbbc70367228403374cb81d33ef109d"} Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.312789 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="580d7626ee9f22f258826a6144eb2af80fbbc70367228403374cb81d33ef109d" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.413086 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zqgnn"] Dec 07 09:37:25 crc kubenswrapper[4838]: E1207 09:37:25.413688 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="registry-server" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.413703 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="registry-server" Dec 07 09:37:25 crc kubenswrapper[4838]: E1207 09:37:25.413724 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05413639-6dc7-443c-9300-975b60426883" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.413731 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="05413639-6dc7-443c-9300-975b60426883" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:37:25 crc kubenswrapper[4838]: E1207 09:37:25.413755 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="extract-utilities" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.413761 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="extract-utilities" Dec 07 09:37:25 crc kubenswrapper[4838]: E1207 09:37:25.413774 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="extract-content" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.413781 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="extract-content" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.413945 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" containerName="registry-server" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.413958 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="05413639-6dc7-443c-9300-975b60426883" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.414503 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.416454 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.417535 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.418066 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.419027 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.433158 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zqgnn"] Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.567360 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.567495 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27fgn\" (UniqueName: \"kubernetes.io/projected/e77e2180-8684-4fbd-8b4c-62d07f058ae8-kube-api-access-27fgn\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.567609 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.630302 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea19715-76c3-4ed3-8e41-ab30f222a3ac" path="/var/lib/kubelet/pods/8ea19715-76c3-4ed3-8e41-ab30f222a3ac/volumes" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.670429 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27fgn\" (UniqueName: \"kubernetes.io/projected/e77e2180-8684-4fbd-8b4c-62d07f058ae8-kube-api-access-27fgn\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.671022 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.671338 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.676593 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.679563 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.694177 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27fgn\" (UniqueName: \"kubernetes.io/projected/e77e2180-8684-4fbd-8b4c-62d07f058ae8-kube-api-access-27fgn\") pod \"ssh-known-hosts-edpm-deployment-zqgnn\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:25 crc kubenswrapper[4838]: I1207 09:37:25.737521 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:26 crc kubenswrapper[4838]: I1207 09:37:26.301096 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zqgnn"] Dec 07 09:37:26 crc kubenswrapper[4838]: I1207 09:37:26.324634 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" event={"ID":"e77e2180-8684-4fbd-8b4c-62d07f058ae8","Type":"ContainerStarted","Data":"f5ca702f778bb24e8cc8409c7d5b430524c718a073de426a58afceec54b3ce86"} Dec 07 09:37:27 crc kubenswrapper[4838]: I1207 09:37:27.341403 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" event={"ID":"e77e2180-8684-4fbd-8b4c-62d07f058ae8","Type":"ContainerStarted","Data":"faba245eb60d21eacae0f9eaef532b5e0d4df4c6bdde2d35404659cbfd4ee0c4"} Dec 07 09:37:27 crc kubenswrapper[4838]: I1207 09:37:27.381140 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" podStartSLOduration=1.912806433 podStartE2EDuration="2.381122414s" podCreationTimestamp="2025-12-07 09:37:25 +0000 UTC" firstStartedPulling="2025-12-07 09:37:26.316260396 +0000 UTC m=+1863.023579453" lastFinishedPulling="2025-12-07 09:37:26.784576417 +0000 UTC m=+1863.491895434" observedRunningTime="2025-12-07 09:37:27.374319717 +0000 UTC m=+1864.081638744" watchObservedRunningTime="2025-12-07 09:37:27.381122414 +0000 UTC m=+1864.088441431" Dec 07 09:37:34 crc kubenswrapper[4838]: I1207 09:37:34.401612 4838 generic.go:334] "Generic (PLEG): container finished" podID="e77e2180-8684-4fbd-8b4c-62d07f058ae8" containerID="faba245eb60d21eacae0f9eaef532b5e0d4df4c6bdde2d35404659cbfd4ee0c4" exitCode=0 Dec 07 09:37:34 crc kubenswrapper[4838]: I1207 09:37:34.401733 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" event={"ID":"e77e2180-8684-4fbd-8b4c-62d07f058ae8","Type":"ContainerDied","Data":"faba245eb60d21eacae0f9eaef532b5e0d4df4c6bdde2d35404659cbfd4ee0c4"} Dec 07 09:37:35 crc kubenswrapper[4838]: I1207 09:37:35.802804 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:35 crc kubenswrapper[4838]: I1207 09:37:35.905386 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27fgn\" (UniqueName: \"kubernetes.io/projected/e77e2180-8684-4fbd-8b4c-62d07f058ae8-kube-api-access-27fgn\") pod \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " Dec 07 09:37:35 crc kubenswrapper[4838]: I1207 09:37:35.905458 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-ssh-key-openstack-edpm-ipam\") pod \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " Dec 07 09:37:35 crc kubenswrapper[4838]: I1207 09:37:35.905801 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-inventory-0\") pod \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\" (UID: \"e77e2180-8684-4fbd-8b4c-62d07f058ae8\") " Dec 07 09:37:35 crc kubenswrapper[4838]: I1207 09:37:35.911551 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e77e2180-8684-4fbd-8b4c-62d07f058ae8-kube-api-access-27fgn" (OuterVolumeSpecName: "kube-api-access-27fgn") pod "e77e2180-8684-4fbd-8b4c-62d07f058ae8" (UID: "e77e2180-8684-4fbd-8b4c-62d07f058ae8"). InnerVolumeSpecName "kube-api-access-27fgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:37:35 crc kubenswrapper[4838]: I1207 09:37:35.929106 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e77e2180-8684-4fbd-8b4c-62d07f058ae8" (UID: "e77e2180-8684-4fbd-8b4c-62d07f058ae8"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:37:35 crc kubenswrapper[4838]: I1207 09:37:35.955729 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e77e2180-8684-4fbd-8b4c-62d07f058ae8" (UID: "e77e2180-8684-4fbd-8b4c-62d07f058ae8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.008104 4838 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.008140 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27fgn\" (UniqueName: \"kubernetes.io/projected/e77e2180-8684-4fbd-8b4c-62d07f058ae8-kube-api-access-27fgn\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.008152 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77e2180-8684-4fbd-8b4c-62d07f058ae8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.421244 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" event={"ID":"e77e2180-8684-4fbd-8b4c-62d07f058ae8","Type":"ContainerDied","Data":"f5ca702f778bb24e8cc8409c7d5b430524c718a073de426a58afceec54b3ce86"} Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.421681 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5ca702f778bb24e8cc8409c7d5b430524c718a073de426a58afceec54b3ce86" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.421342 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zqgnn" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.523765 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm"] Dec 07 09:37:36 crc kubenswrapper[4838]: E1207 09:37:36.524255 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77e2180-8684-4fbd-8b4c-62d07f058ae8" containerName="ssh-known-hosts-edpm-deployment" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.524279 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77e2180-8684-4fbd-8b4c-62d07f058ae8" containerName="ssh-known-hosts-edpm-deployment" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.524520 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e77e2180-8684-4fbd-8b4c-62d07f058ae8" containerName="ssh-known-hosts-edpm-deployment" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.525339 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.529224 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.529565 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.529717 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.531424 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.540268 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm"] Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.617974 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.618057 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjtws\" (UniqueName: \"kubernetes.io/projected/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-kube-api-access-vjtws\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.618142 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.720095 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjtws\" (UniqueName: \"kubernetes.io/projected/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-kube-api-access-vjtws\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.720218 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.720308 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.726780 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.728095 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.739211 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjtws\" (UniqueName: \"kubernetes.io/projected/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-kube-api-access-vjtws\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4jjzm\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:36 crc kubenswrapper[4838]: I1207 09:37:36.901328 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:37 crc kubenswrapper[4838]: I1207 09:37:37.399974 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm"] Dec 07 09:37:37 crc kubenswrapper[4838]: I1207 09:37:37.433243 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" event={"ID":"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e","Type":"ContainerStarted","Data":"305154fdf20a2b6bc1f0f3c78b0f0164557d1bb8d702c270a9d456adb41aebca"} Dec 07 09:37:38 crc kubenswrapper[4838]: I1207 09:37:38.443622 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" event={"ID":"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e","Type":"ContainerStarted","Data":"5eb0970e21d534da8fa970ee8241ea8e67ae26d54d9e78aa3b69595227fd1e55"} Dec 07 09:37:38 crc kubenswrapper[4838]: I1207 09:37:38.459842 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" podStartSLOduration=2.017564932 podStartE2EDuration="2.45980885s" podCreationTimestamp="2025-12-07 09:37:36 +0000 UTC" firstStartedPulling="2025-12-07 09:37:37.413991643 +0000 UTC m=+1874.121310670" lastFinishedPulling="2025-12-07 09:37:37.856235571 +0000 UTC m=+1874.563554588" observedRunningTime="2025-12-07 09:37:38.458917395 +0000 UTC m=+1875.166236432" watchObservedRunningTime="2025-12-07 09:37:38.45980885 +0000 UTC m=+1875.167127867" Dec 07 09:37:43 crc kubenswrapper[4838]: I1207 09:37:43.058365 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h7vd6"] Dec 07 09:37:43 crc kubenswrapper[4838]: I1207 09:37:43.069704 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h7vd6"] Dec 07 09:37:43 crc kubenswrapper[4838]: I1207 09:37:43.079186 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-v59wv"] Dec 07 09:37:43 crc kubenswrapper[4838]: I1207 09:37:43.087730 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-v59wv"] Dec 07 09:37:43 crc kubenswrapper[4838]: I1207 09:37:43.627078 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63fb1e51-5812-4adf-9e44-a7f4cd3349da" path="/var/lib/kubelet/pods/63fb1e51-5812-4adf-9e44-a7f4cd3349da/volumes" Dec 07 09:37:43 crc kubenswrapper[4838]: I1207 09:37:43.627580 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b4af2f-6fae-4a99-a9ee-8447c2a00df3" path="/var/lib/kubelet/pods/99b4af2f-6fae-4a99-a9ee-8447c2a00df3/volumes" Dec 07 09:37:47 crc kubenswrapper[4838]: I1207 09:37:47.567833 4838 generic.go:334] "Generic (PLEG): container finished" podID="7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" containerID="5eb0970e21d534da8fa970ee8241ea8e67ae26d54d9e78aa3b69595227fd1e55" exitCode=0 Dec 07 09:37:47 crc kubenswrapper[4838]: I1207 09:37:47.567878 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" event={"ID":"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e","Type":"ContainerDied","Data":"5eb0970e21d534da8fa970ee8241ea8e67ae26d54d9e78aa3b69595227fd1e55"} Dec 07 09:37:48 crc kubenswrapper[4838]: I1207 09:37:48.979018 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.077699 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-ssh-key\") pod \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.078024 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-inventory\") pod \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.078189 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjtws\" (UniqueName: \"kubernetes.io/projected/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-kube-api-access-vjtws\") pod \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\" (UID: \"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e\") " Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.091869 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-kube-api-access-vjtws" (OuterVolumeSpecName: "kube-api-access-vjtws") pod "7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" (UID: "7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e"). InnerVolumeSpecName "kube-api-access-vjtws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.104175 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-inventory" (OuterVolumeSpecName: "inventory") pod "7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" (UID: "7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.124677 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" (UID: "7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.180911 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.181279 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.181294 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjtws\" (UniqueName: \"kubernetes.io/projected/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e-kube-api-access-vjtws\") on node \"crc\" DevicePath \"\"" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.591093 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" event={"ID":"7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e","Type":"ContainerDied","Data":"305154fdf20a2b6bc1f0f3c78b0f0164557d1bb8d702c270a9d456adb41aebca"} Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.591152 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="305154fdf20a2b6bc1f0f3c78b0f0164557d1bb8d702c270a9d456adb41aebca" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.591205 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.686437 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4"] Dec 07 09:37:49 crc kubenswrapper[4838]: E1207 09:37:49.686865 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.686884 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.687133 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.687993 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.692593 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.692963 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.693335 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.693471 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.701305 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4"] Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.792665 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.792950 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2vd8\" (UniqueName: \"kubernetes.io/projected/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-kube-api-access-c2vd8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.793063 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.894360 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.894486 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2vd8\" (UniqueName: \"kubernetes.io/projected/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-kube-api-access-c2vd8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.894525 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.898835 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.898886 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:49 crc kubenswrapper[4838]: I1207 09:37:49.915204 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2vd8\" (UniqueName: \"kubernetes.io/projected/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-kube-api-access-c2vd8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:50 crc kubenswrapper[4838]: I1207 09:37:50.015250 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:37:50 crc kubenswrapper[4838]: I1207 09:37:50.563352 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4"] Dec 07 09:37:50 crc kubenswrapper[4838]: I1207 09:37:50.602845 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" event={"ID":"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297","Type":"ContainerStarted","Data":"c3c12c6e3f44f29e705074231661ce2b4cd6e95d3eb33fbc6d48dba3d3ff8813"} Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.295611 4838 scope.go:117] "RemoveContainer" containerID="f1331b139ed990a28aebd176d4684b6187ee3235742df54bf6b28726119d4445" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.335485 4838 scope.go:117] "RemoveContainer" containerID="5579b735ad3e68795d3aaf69421e09ac15e0ff5575a5d4fc49b15be5f531e2ad" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.368178 4838 scope.go:117] "RemoveContainer" containerID="7c5b1ddc9b76a5884073df34fc8a662a7c9011f864542fc08d94232f96a7b787" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.417490 4838 scope.go:117] "RemoveContainer" containerID="ea9efd8dc3bbbbbd175f07376d3f441b917ce8f4d6002a9071bf2234357e05d7" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.457607 4838 scope.go:117] "RemoveContainer" containerID="0763ab69bbf4838c83b1557c1944716eacb72716e08d734380faf656afae2b9e" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.477373 4838 scope.go:117] "RemoveContainer" containerID="c7ec1855907e09fdb26e3e71384c14e52427ce6c0a56f97408e3fe2c5073fb7d" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.499602 4838 scope.go:117] "RemoveContainer" containerID="7ddff9e775deed21a1736b371cbedd551de97ea5d9f8105586db67c68865d63c" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.541056 4838 scope.go:117] "RemoveContainer" containerID="9f65f736b1524df7292ee3e57c12ec4628e3c67c76e5d01ab4788811b0e44b6f" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.559079 4838 scope.go:117] "RemoveContainer" containerID="4eb99b3245d0e86a5b8360cd9b492896486e98c4d81604616e45226be633955c" Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.628562 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" event={"ID":"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297","Type":"ContainerStarted","Data":"a5a982b3242fa6920d9f10ca35ad81cc37c5c4395a66cf9a1cc73ffbd87e19d1"} Dec 07 09:37:51 crc kubenswrapper[4838]: I1207 09:37:51.634543 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" podStartSLOduration=2.228051205 podStartE2EDuration="2.634525062s" podCreationTimestamp="2025-12-07 09:37:49 +0000 UTC" firstStartedPulling="2025-12-07 09:37:50.567710447 +0000 UTC m=+1887.275029464" lastFinishedPulling="2025-12-07 09:37:50.974184284 +0000 UTC m=+1887.681503321" observedRunningTime="2025-12-07 09:37:51.632318568 +0000 UTC m=+1888.339637585" watchObservedRunningTime="2025-12-07 09:37:51.634525062 +0000 UTC m=+1888.341844079" Dec 07 09:38:02 crc kubenswrapper[4838]: I1207 09:38:02.747056 4838 generic.go:334] "Generic (PLEG): container finished" podID="1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" containerID="a5a982b3242fa6920d9f10ca35ad81cc37c5c4395a66cf9a1cc73ffbd87e19d1" exitCode=0 Dec 07 09:38:02 crc kubenswrapper[4838]: I1207 09:38:02.747157 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" event={"ID":"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297","Type":"ContainerDied","Data":"a5a982b3242fa6920d9f10ca35ad81cc37c5c4395a66cf9a1cc73ffbd87e19d1"} Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.179198 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.278618 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2vd8\" (UniqueName: \"kubernetes.io/projected/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-kube-api-access-c2vd8\") pod \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.278741 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-inventory\") pod \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.278912 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-ssh-key\") pod \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\" (UID: \"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297\") " Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.284854 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-kube-api-access-c2vd8" (OuterVolumeSpecName: "kube-api-access-c2vd8") pod "1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" (UID: "1b64d9f5-b1db-4ebe-893e-c8da4e2d0297"). InnerVolumeSpecName "kube-api-access-c2vd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.313752 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" (UID: "1b64d9f5-b1db-4ebe-893e-c8da4e2d0297"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.317855 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-inventory" (OuterVolumeSpecName: "inventory") pod "1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" (UID: "1b64d9f5-b1db-4ebe-893e-c8da4e2d0297"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.382850 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2vd8\" (UniqueName: \"kubernetes.io/projected/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-kube-api-access-c2vd8\") on node \"crc\" DevicePath \"\"" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.382892 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.382907 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.774599 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" event={"ID":"1b64d9f5-b1db-4ebe-893e-c8da4e2d0297","Type":"ContainerDied","Data":"c3c12c6e3f44f29e705074231661ce2b4cd6e95d3eb33fbc6d48dba3d3ff8813"} Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.774662 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3c12c6e3f44f29e705074231661ce2b4cd6e95d3eb33fbc6d48dba3d3ff8813" Dec 07 09:38:04 crc kubenswrapper[4838]: I1207 09:38:04.775376 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4" Dec 07 09:38:26 crc kubenswrapper[4838]: I1207 09:38:26.061134 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-wndph"] Dec 07 09:38:26 crc kubenswrapper[4838]: I1207 09:38:26.072300 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-wndph"] Dec 07 09:38:27 crc kubenswrapper[4838]: I1207 09:38:27.629655 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9fd40d9-0ac7-47d6-afd7-05a6a88ef137" path="/var/lib/kubelet/pods/e9fd40d9-0ac7-47d6-afd7-05a6a88ef137/volumes" Dec 07 09:38:51 crc kubenswrapper[4838]: I1207 09:38:51.838530 4838 scope.go:117] "RemoveContainer" containerID="b25b83605bcbe9e2002ca0fe388109c280cc08dd827f6dc0b00c740c1151d4ca" Dec 07 09:39:24 crc kubenswrapper[4838]: I1207 09:39:24.493232 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:39:24 crc kubenswrapper[4838]: I1207 09:39:24.494051 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.076452 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gdnvr"] Dec 07 09:39:51 crc kubenswrapper[4838]: E1207 09:39:51.077364 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.077378 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.077558 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.078823 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.108063 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdnvr"] Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.158312 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-catalog-content\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.158409 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffgh8\" (UniqueName: \"kubernetes.io/projected/50505154-c1c9-4d95-b906-b02bcc56579d-kube-api-access-ffgh8\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.158483 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-utilities\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.259955 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffgh8\" (UniqueName: \"kubernetes.io/projected/50505154-c1c9-4d95-b906-b02bcc56579d-kube-api-access-ffgh8\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.260064 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-utilities\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.260127 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-catalog-content\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.260625 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-utilities\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.260714 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-catalog-content\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.286316 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffgh8\" (UniqueName: \"kubernetes.io/projected/50505154-c1c9-4d95-b906-b02bcc56579d-kube-api-access-ffgh8\") pod \"redhat-operators-gdnvr\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.409485 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:39:51 crc kubenswrapper[4838]: I1207 09:39:51.997451 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdnvr"] Dec 07 09:39:52 crc kubenswrapper[4838]: W1207 09:39:52.007567 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50505154_c1c9_4d95_b906_b02bcc56579d.slice/crio-bb5cebce67f2607ef7a0eb218aeab5521f2e6a0e1eca8d8fcc2f9059d60df1c7 WatchSource:0}: Error finding container bb5cebce67f2607ef7a0eb218aeab5521f2e6a0e1eca8d8fcc2f9059d60df1c7: Status 404 returned error can't find the container with id bb5cebce67f2607ef7a0eb218aeab5521f2e6a0e1eca8d8fcc2f9059d60df1c7 Dec 07 09:39:52 crc kubenswrapper[4838]: I1207 09:39:52.788712 4838 generic.go:334] "Generic (PLEG): container finished" podID="50505154-c1c9-4d95-b906-b02bcc56579d" containerID="b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea" exitCode=0 Dec 07 09:39:52 crc kubenswrapper[4838]: I1207 09:39:52.788778 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdnvr" event={"ID":"50505154-c1c9-4d95-b906-b02bcc56579d","Type":"ContainerDied","Data":"b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea"} Dec 07 09:39:52 crc kubenswrapper[4838]: I1207 09:39:52.789046 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdnvr" event={"ID":"50505154-c1c9-4d95-b906-b02bcc56579d","Type":"ContainerStarted","Data":"bb5cebce67f2607ef7a0eb218aeab5521f2e6a0e1eca8d8fcc2f9059d60df1c7"} Dec 07 09:39:52 crc kubenswrapper[4838]: I1207 09:39:52.791226 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:39:53 crc kubenswrapper[4838]: I1207 09:39:53.801728 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdnvr" event={"ID":"50505154-c1c9-4d95-b906-b02bcc56579d","Type":"ContainerStarted","Data":"96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce"} Dec 07 09:39:54 crc kubenswrapper[4838]: I1207 09:39:54.494310 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:39:54 crc kubenswrapper[4838]: I1207 09:39:54.494384 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:39:57 crc kubenswrapper[4838]: I1207 09:39:57.840567 4838 generic.go:334] "Generic (PLEG): container finished" podID="50505154-c1c9-4d95-b906-b02bcc56579d" containerID="96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce" exitCode=0 Dec 07 09:39:57 crc kubenswrapper[4838]: I1207 09:39:57.840655 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdnvr" event={"ID":"50505154-c1c9-4d95-b906-b02bcc56579d","Type":"ContainerDied","Data":"96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce"} Dec 07 09:39:58 crc kubenswrapper[4838]: I1207 09:39:58.850086 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdnvr" event={"ID":"50505154-c1c9-4d95-b906-b02bcc56579d","Type":"ContainerStarted","Data":"077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d"} Dec 07 09:39:58 crc kubenswrapper[4838]: I1207 09:39:58.911001 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gdnvr" podStartSLOduration=2.416728687 podStartE2EDuration="7.910961275s" podCreationTimestamp="2025-12-07 09:39:51 +0000 UTC" firstStartedPulling="2025-12-07 09:39:52.79103068 +0000 UTC m=+2009.498349697" lastFinishedPulling="2025-12-07 09:39:58.285263268 +0000 UTC m=+2014.992582285" observedRunningTime="2025-12-07 09:39:58.8774315 +0000 UTC m=+2015.584750517" watchObservedRunningTime="2025-12-07 09:39:58.910961275 +0000 UTC m=+2015.618280292" Dec 07 09:40:01 crc kubenswrapper[4838]: I1207 09:40:01.410529 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:40:01 crc kubenswrapper[4838]: I1207 09:40:01.410584 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:40:02 crc kubenswrapper[4838]: I1207 09:40:02.452360 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdnvr" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="registry-server" probeResult="failure" output=< Dec 07 09:40:02 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 09:40:02 crc kubenswrapper[4838]: > Dec 07 09:40:11 crc kubenswrapper[4838]: I1207 09:40:11.455737 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:40:11 crc kubenswrapper[4838]: I1207 09:40:11.504589 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:40:11 crc kubenswrapper[4838]: I1207 09:40:11.691123 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdnvr"] Dec 07 09:40:12 crc kubenswrapper[4838]: I1207 09:40:12.960740 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gdnvr" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="registry-server" containerID="cri-o://077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d" gracePeriod=2 Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.464278 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.522453 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffgh8\" (UniqueName: \"kubernetes.io/projected/50505154-c1c9-4d95-b906-b02bcc56579d-kube-api-access-ffgh8\") pod \"50505154-c1c9-4d95-b906-b02bcc56579d\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.523105 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-catalog-content\") pod \"50505154-c1c9-4d95-b906-b02bcc56579d\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.531028 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-utilities\") pod \"50505154-c1c9-4d95-b906-b02bcc56579d\" (UID: \"50505154-c1c9-4d95-b906-b02bcc56579d\") " Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.532306 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-utilities" (OuterVolumeSpecName: "utilities") pod "50505154-c1c9-4d95-b906-b02bcc56579d" (UID: "50505154-c1c9-4d95-b906-b02bcc56579d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.552354 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50505154-c1c9-4d95-b906-b02bcc56579d-kube-api-access-ffgh8" (OuterVolumeSpecName: "kube-api-access-ffgh8") pod "50505154-c1c9-4d95-b906-b02bcc56579d" (UID: "50505154-c1c9-4d95-b906-b02bcc56579d"). InnerVolumeSpecName "kube-api-access-ffgh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.633220 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffgh8\" (UniqueName: \"kubernetes.io/projected/50505154-c1c9-4d95-b906-b02bcc56579d-kube-api-access-ffgh8\") on node \"crc\" DevicePath \"\"" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.633254 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.635790 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50505154-c1c9-4d95-b906-b02bcc56579d" (UID: "50505154-c1c9-4d95-b906-b02bcc56579d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.736922 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50505154-c1c9-4d95-b906-b02bcc56579d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.969780 4838 generic.go:334] "Generic (PLEG): container finished" podID="50505154-c1c9-4d95-b906-b02bcc56579d" containerID="077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d" exitCode=0 Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.969844 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdnvr" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.969868 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdnvr" event={"ID":"50505154-c1c9-4d95-b906-b02bcc56579d","Type":"ContainerDied","Data":"077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d"} Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.970689 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdnvr" event={"ID":"50505154-c1c9-4d95-b906-b02bcc56579d","Type":"ContainerDied","Data":"bb5cebce67f2607ef7a0eb218aeab5521f2e6a0e1eca8d8fcc2f9059d60df1c7"} Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.970711 4838 scope.go:117] "RemoveContainer" containerID="077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d" Dec 07 09:40:13 crc kubenswrapper[4838]: I1207 09:40:13.990774 4838 scope.go:117] "RemoveContainer" containerID="96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce" Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.004497 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdnvr"] Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.013646 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gdnvr"] Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.014037 4838 scope.go:117] "RemoveContainer" containerID="b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea" Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.043960 4838 scope.go:117] "RemoveContainer" containerID="077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d" Dec 07 09:40:14 crc kubenswrapper[4838]: E1207 09:40:14.044456 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d\": container with ID starting with 077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d not found: ID does not exist" containerID="077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d" Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.044492 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d"} err="failed to get container status \"077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d\": rpc error: code = NotFound desc = could not find container \"077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d\": container with ID starting with 077dfb62be1ade5d073db444cf1dcb67a81befe3f4ca0d926685ff3df881000d not found: ID does not exist" Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.044515 4838 scope.go:117] "RemoveContainer" containerID="96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce" Dec 07 09:40:14 crc kubenswrapper[4838]: E1207 09:40:14.044913 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce\": container with ID starting with 96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce not found: ID does not exist" containerID="96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce" Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.044976 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce"} err="failed to get container status \"96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce\": rpc error: code = NotFound desc = could not find container \"96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce\": container with ID starting with 96fc7e7c5086b0f90ace1645ba1c3bc32532fe580b99ff4ebc4b275d8a1328ce not found: ID does not exist" Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.045007 4838 scope.go:117] "RemoveContainer" containerID="b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea" Dec 07 09:40:14 crc kubenswrapper[4838]: E1207 09:40:14.045414 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea\": container with ID starting with b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea not found: ID does not exist" containerID="b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea" Dec 07 09:40:14 crc kubenswrapper[4838]: I1207 09:40:14.045455 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea"} err="failed to get container status \"b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea\": rpc error: code = NotFound desc = could not find container \"b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea\": container with ID starting with b2fd80eac97d4697f40cb72563a550b7ca8e0131dd9519004cc7b3cb329e7aea not found: ID does not exist" Dec 07 09:40:15 crc kubenswrapper[4838]: I1207 09:40:15.625640 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" path="/var/lib/kubelet/pods/50505154-c1c9-4d95-b906-b02bcc56579d/volumes" Dec 07 09:40:24 crc kubenswrapper[4838]: I1207 09:40:24.493399 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:40:24 crc kubenswrapper[4838]: I1207 09:40:24.494214 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:40:24 crc kubenswrapper[4838]: I1207 09:40:24.494273 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:40:24 crc kubenswrapper[4838]: I1207 09:40:24.495089 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"81353be09f7749f5b2e22eca0a163e7ec79346eb5cbc71ad52479317f9293c04"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:40:24 crc kubenswrapper[4838]: I1207 09:40:24.495172 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://81353be09f7749f5b2e22eca0a163e7ec79346eb5cbc71ad52479317f9293c04" gracePeriod=600 Dec 07 09:40:25 crc kubenswrapper[4838]: I1207 09:40:25.057400 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="81353be09f7749f5b2e22eca0a163e7ec79346eb5cbc71ad52479317f9293c04" exitCode=0 Dec 07 09:40:25 crc kubenswrapper[4838]: I1207 09:40:25.057669 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"81353be09f7749f5b2e22eca0a163e7ec79346eb5cbc71ad52479317f9293c04"} Dec 07 09:40:25 crc kubenswrapper[4838]: I1207 09:40:25.057695 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8"} Dec 07 09:40:25 crc kubenswrapper[4838]: I1207 09:40:25.057711 4838 scope.go:117] "RemoveContainer" containerID="7446a74a52c115ab068e8606d145db8971f8df77c7ba0f2530321a99bf5b12c3" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.714300 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h27t4"] Dec 07 09:42:08 crc kubenswrapper[4838]: E1207 09:42:08.715237 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="registry-server" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.715250 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="registry-server" Dec 07 09:42:08 crc kubenswrapper[4838]: E1207 09:42:08.715266 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="extract-utilities" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.715272 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="extract-utilities" Dec 07 09:42:08 crc kubenswrapper[4838]: E1207 09:42:08.715278 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="extract-content" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.715285 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="extract-content" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.715448 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="50505154-c1c9-4d95-b906-b02bcc56579d" containerName="registry-server" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.716639 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.739540 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h27t4"] Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.820460 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-utilities\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.820709 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-catalog-content\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.820781 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjlzp\" (UniqueName: \"kubernetes.io/projected/cd723953-667b-4948-948b-d4f1ad03f021-kube-api-access-qjlzp\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.909955 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hp8b4"] Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.911665 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.923883 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-utilities\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.924075 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-catalog-content\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.924127 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjlzp\" (UniqueName: \"kubernetes.io/projected/cd723953-667b-4948-948b-d4f1ad03f021-kube-api-access-qjlzp\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.924915 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-utilities\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.925215 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-catalog-content\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.925243 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hp8b4"] Dec 07 09:42:08 crc kubenswrapper[4838]: I1207 09:42:08.968935 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjlzp\" (UniqueName: \"kubernetes.io/projected/cd723953-667b-4948-948b-d4f1ad03f021-kube-api-access-qjlzp\") pod \"community-operators-h27t4\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.025860 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-utilities\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.026089 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-catalog-content\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.026222 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76xtt\" (UniqueName: \"kubernetes.io/projected/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-kube-api-access-76xtt\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.041975 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.127696 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-utilities\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.128112 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-catalog-content\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.128153 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76xtt\" (UniqueName: \"kubernetes.io/projected/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-kube-api-access-76xtt\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.128158 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-utilities\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.128883 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-catalog-content\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.154449 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76xtt\" (UniqueName: \"kubernetes.io/projected/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-kube-api-access-76xtt\") pod \"certified-operators-hp8b4\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.226613 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.713878 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h27t4"] Dec 07 09:42:09 crc kubenswrapper[4838]: I1207 09:42:09.895191 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hp8b4"] Dec 07 09:42:10 crc kubenswrapper[4838]: I1207 09:42:10.014257 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h27t4" event={"ID":"cd723953-667b-4948-948b-d4f1ad03f021","Type":"ContainerStarted","Data":"cf2f69f6984d9f022df21a827c23750abb258d00137b4867d63a7f3bb163d10d"} Dec 07 09:42:10 crc kubenswrapper[4838]: I1207 09:42:10.023312 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hp8b4" event={"ID":"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5","Type":"ContainerStarted","Data":"a0febadac1ce4471bc3cca7e6bbf96dce64eee690ad3a1d643f0c90ec50858e5"} Dec 07 09:42:11 crc kubenswrapper[4838]: I1207 09:42:11.035403 4838 generic.go:334] "Generic (PLEG): container finished" podID="cd723953-667b-4948-948b-d4f1ad03f021" containerID="8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f" exitCode=0 Dec 07 09:42:11 crc kubenswrapper[4838]: I1207 09:42:11.035493 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h27t4" event={"ID":"cd723953-667b-4948-948b-d4f1ad03f021","Type":"ContainerDied","Data":"8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f"} Dec 07 09:42:11 crc kubenswrapper[4838]: I1207 09:42:11.040841 4838 generic.go:334] "Generic (PLEG): container finished" podID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerID="ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a" exitCode=0 Dec 07 09:42:11 crc kubenswrapper[4838]: I1207 09:42:11.040888 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hp8b4" event={"ID":"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5","Type":"ContainerDied","Data":"ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a"} Dec 07 09:42:12 crc kubenswrapper[4838]: I1207 09:42:12.055806 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hp8b4" event={"ID":"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5","Type":"ContainerStarted","Data":"b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d"} Dec 07 09:42:13 crc kubenswrapper[4838]: I1207 09:42:13.068981 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h27t4" event={"ID":"cd723953-667b-4948-948b-d4f1ad03f021","Type":"ContainerStarted","Data":"e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050"} Dec 07 09:42:15 crc kubenswrapper[4838]: I1207 09:42:15.093667 4838 generic.go:334] "Generic (PLEG): container finished" podID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerID="b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d" exitCode=0 Dec 07 09:42:15 crc kubenswrapper[4838]: I1207 09:42:15.093730 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hp8b4" event={"ID":"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5","Type":"ContainerDied","Data":"b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d"} Dec 07 09:42:15 crc kubenswrapper[4838]: I1207 09:42:15.096524 4838 generic.go:334] "Generic (PLEG): container finished" podID="cd723953-667b-4948-948b-d4f1ad03f021" containerID="e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050" exitCode=0 Dec 07 09:42:15 crc kubenswrapper[4838]: I1207 09:42:15.096564 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h27t4" event={"ID":"cd723953-667b-4948-948b-d4f1ad03f021","Type":"ContainerDied","Data":"e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050"} Dec 07 09:42:16 crc kubenswrapper[4838]: I1207 09:42:16.114505 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hp8b4" event={"ID":"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5","Type":"ContainerStarted","Data":"573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896"} Dec 07 09:42:16 crc kubenswrapper[4838]: I1207 09:42:16.117747 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h27t4" event={"ID":"cd723953-667b-4948-948b-d4f1ad03f021","Type":"ContainerStarted","Data":"d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8"} Dec 07 09:42:16 crc kubenswrapper[4838]: I1207 09:42:16.140626 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hp8b4" podStartSLOduration=3.652016044 podStartE2EDuration="8.140606666s" podCreationTimestamp="2025-12-07 09:42:08 +0000 UTC" firstStartedPulling="2025-12-07 09:42:11.043121646 +0000 UTC m=+2147.750440703" lastFinishedPulling="2025-12-07 09:42:15.531712308 +0000 UTC m=+2152.239031325" observedRunningTime="2025-12-07 09:42:16.1362829 +0000 UTC m=+2152.843601927" watchObservedRunningTime="2025-12-07 09:42:16.140606666 +0000 UTC m=+2152.847925683" Dec 07 09:42:16 crc kubenswrapper[4838]: I1207 09:42:16.164228 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h27t4" podStartSLOduration=3.735358203 podStartE2EDuration="8.164200087s" podCreationTimestamp="2025-12-07 09:42:08 +0000 UTC" firstStartedPulling="2025-12-07 09:42:11.039948033 +0000 UTC m=+2147.747267050" lastFinishedPulling="2025-12-07 09:42:15.468789917 +0000 UTC m=+2152.176108934" observedRunningTime="2025-12-07 09:42:16.156226274 +0000 UTC m=+2152.863545291" watchObservedRunningTime="2025-12-07 09:42:16.164200087 +0000 UTC m=+2152.871519104" Dec 07 09:42:19 crc kubenswrapper[4838]: I1207 09:42:19.043179 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:19 crc kubenswrapper[4838]: I1207 09:42:19.044600 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:19 crc kubenswrapper[4838]: I1207 09:42:19.097293 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:19 crc kubenswrapper[4838]: I1207 09:42:19.227570 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:19 crc kubenswrapper[4838]: I1207 09:42:19.227808 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:19 crc kubenswrapper[4838]: I1207 09:42:19.278308 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:20 crc kubenswrapper[4838]: I1207 09:42:20.198243 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:21 crc kubenswrapper[4838]: I1207 09:42:21.302748 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hp8b4"] Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.171264 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hp8b4" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="registry-server" containerID="cri-o://573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896" gracePeriod=2 Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.631620 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.700118 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-catalog-content\") pod \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.700189 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76xtt\" (UniqueName: \"kubernetes.io/projected/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-kube-api-access-76xtt\") pod \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.700311 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-utilities\") pod \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\" (UID: \"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5\") " Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.705702 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-utilities" (OuterVolumeSpecName: "utilities") pod "6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" (UID: "6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.711427 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-kube-api-access-76xtt" (OuterVolumeSpecName: "kube-api-access-76xtt") pod "6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" (UID: "6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5"). InnerVolumeSpecName "kube-api-access-76xtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.763106 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" (UID: "6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.802119 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76xtt\" (UniqueName: \"kubernetes.io/projected/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-kube-api-access-76xtt\") on node \"crc\" DevicePath \"\"" Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.802344 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:42:23 crc kubenswrapper[4838]: I1207 09:42:23.802436 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.180670 4838 generic.go:334] "Generic (PLEG): container finished" podID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerID="573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896" exitCode=0 Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.180718 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hp8b4" event={"ID":"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5","Type":"ContainerDied","Data":"573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896"} Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.180735 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hp8b4" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.180751 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hp8b4" event={"ID":"6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5","Type":"ContainerDied","Data":"a0febadac1ce4471bc3cca7e6bbf96dce64eee690ad3a1d643f0c90ec50858e5"} Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.180774 4838 scope.go:117] "RemoveContainer" containerID="573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.226195 4838 scope.go:117] "RemoveContainer" containerID="b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.230846 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hp8b4"] Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.241404 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hp8b4"] Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.249174 4838 scope.go:117] "RemoveContainer" containerID="ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.298378 4838 scope.go:117] "RemoveContainer" containerID="573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896" Dec 07 09:42:24 crc kubenswrapper[4838]: E1207 09:42:24.298847 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896\": container with ID starting with 573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896 not found: ID does not exist" containerID="573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.298874 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896"} err="failed to get container status \"573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896\": rpc error: code = NotFound desc = could not find container \"573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896\": container with ID starting with 573f93d1ad8fd0a40eb9d7a97afdcece309b0a7341542b17d787290394b13896 not found: ID does not exist" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.298894 4838 scope.go:117] "RemoveContainer" containerID="b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d" Dec 07 09:42:24 crc kubenswrapper[4838]: E1207 09:42:24.299287 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d\": container with ID starting with b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d not found: ID does not exist" containerID="b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.299306 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d"} err="failed to get container status \"b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d\": rpc error: code = NotFound desc = could not find container \"b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d\": container with ID starting with b960e20c5300bca10c7f48baa7bfa701e59dde793c8406fd36d60505113f0d4d not found: ID does not exist" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.299318 4838 scope.go:117] "RemoveContainer" containerID="ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a" Dec 07 09:42:24 crc kubenswrapper[4838]: E1207 09:42:24.299709 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a\": container with ID starting with ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a not found: ID does not exist" containerID="ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.299729 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a"} err="failed to get container status \"ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a\": rpc error: code = NotFound desc = could not find container \"ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a\": container with ID starting with ff7fd8b3b663c2ba9f4ebd8f7cb7cbf84a2f2fc2ef27bc9458195e555a5f9b0a not found: ID does not exist" Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.492572 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:42:24 crc kubenswrapper[4838]: I1207 09:42:24.492629 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:42:25 crc kubenswrapper[4838]: I1207 09:42:25.627756 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" path="/var/lib/kubelet/pods/6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5/volumes" Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.105214 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.154556 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h27t4"] Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.226230 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h27t4" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="registry-server" containerID="cri-o://d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8" gracePeriod=2 Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.856134 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.902511 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-catalog-content\") pod \"cd723953-667b-4948-948b-d4f1ad03f021\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.902613 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-utilities\") pod \"cd723953-667b-4948-948b-d4f1ad03f021\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.902682 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjlzp\" (UniqueName: \"kubernetes.io/projected/cd723953-667b-4948-948b-d4f1ad03f021-kube-api-access-qjlzp\") pod \"cd723953-667b-4948-948b-d4f1ad03f021\" (UID: \"cd723953-667b-4948-948b-d4f1ad03f021\") " Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.903464 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-utilities" (OuterVolumeSpecName: "utilities") pod "cd723953-667b-4948-948b-d4f1ad03f021" (UID: "cd723953-667b-4948-948b-d4f1ad03f021"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.919114 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd723953-667b-4948-948b-d4f1ad03f021-kube-api-access-qjlzp" (OuterVolumeSpecName: "kube-api-access-qjlzp") pod "cd723953-667b-4948-948b-d4f1ad03f021" (UID: "cd723953-667b-4948-948b-d4f1ad03f021"). InnerVolumeSpecName "kube-api-access-qjlzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:42:29 crc kubenswrapper[4838]: I1207 09:42:29.958091 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd723953-667b-4948-948b-d4f1ad03f021" (UID: "cd723953-667b-4948-948b-d4f1ad03f021"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.004468 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.004516 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjlzp\" (UniqueName: \"kubernetes.io/projected/cd723953-667b-4948-948b-d4f1ad03f021-kube-api-access-qjlzp\") on node \"crc\" DevicePath \"\"" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.004527 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd723953-667b-4948-948b-d4f1ad03f021-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.234108 4838 generic.go:334] "Generic (PLEG): container finished" podID="cd723953-667b-4948-948b-d4f1ad03f021" containerID="d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8" exitCode=0 Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.234148 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h27t4" event={"ID":"cd723953-667b-4948-948b-d4f1ad03f021","Type":"ContainerDied","Data":"d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8"} Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.234164 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h27t4" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.234175 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h27t4" event={"ID":"cd723953-667b-4948-948b-d4f1ad03f021","Type":"ContainerDied","Data":"cf2f69f6984d9f022df21a827c23750abb258d00137b4867d63a7f3bb163d10d"} Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.234196 4838 scope.go:117] "RemoveContainer" containerID="d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.252437 4838 scope.go:117] "RemoveContainer" containerID="e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.269593 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h27t4"] Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.274092 4838 scope.go:117] "RemoveContainer" containerID="8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.289222 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h27t4"] Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.314940 4838 scope.go:117] "RemoveContainer" containerID="d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8" Dec 07 09:42:30 crc kubenswrapper[4838]: E1207 09:42:30.315367 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8\": container with ID starting with d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8 not found: ID does not exist" containerID="d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.315396 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8"} err="failed to get container status \"d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8\": rpc error: code = NotFound desc = could not find container \"d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8\": container with ID starting with d6753da77c7c1989e9e8c7a80a17f15ad64a2301e073a4abb2ed67bf8fc5d3d8 not found: ID does not exist" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.315414 4838 scope.go:117] "RemoveContainer" containerID="e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050" Dec 07 09:42:30 crc kubenswrapper[4838]: E1207 09:42:30.315641 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050\": container with ID starting with e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050 not found: ID does not exist" containerID="e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.315665 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050"} err="failed to get container status \"e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050\": rpc error: code = NotFound desc = could not find container \"e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050\": container with ID starting with e984ba07e78459644708a6d562bd1be25025b16d907b1f5799a50238d3a60050 not found: ID does not exist" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.315679 4838 scope.go:117] "RemoveContainer" containerID="8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f" Dec 07 09:42:30 crc kubenswrapper[4838]: E1207 09:42:30.315848 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f\": container with ID starting with 8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f not found: ID does not exist" containerID="8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f" Dec 07 09:42:30 crc kubenswrapper[4838]: I1207 09:42:30.315873 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f"} err="failed to get container status \"8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f\": rpc error: code = NotFound desc = could not find container \"8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f\": container with ID starting with 8b94204dd1c340e2247f13b6e25ab7b67e74a3188b17d0c0cb376e0c1262429f not found: ID does not exist" Dec 07 09:42:31 crc kubenswrapper[4838]: I1207 09:42:31.623968 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd723953-667b-4948-948b-d4f1ad03f021" path="/var/lib/kubelet/pods/cd723953-667b-4948-948b-d4f1ad03f021/volumes" Dec 07 09:42:54 crc kubenswrapper[4838]: I1207 09:42:54.493433 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:42:54 crc kubenswrapper[4838]: I1207 09:42:54.493955 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.493522 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.494063 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.494113 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.494746 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.494789 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" gracePeriod=600 Dec 07 09:43:24 crc kubenswrapper[4838]: E1207 09:43:24.620358 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.748678 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" exitCode=0 Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.748720 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8"} Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.748751 4838 scope.go:117] "RemoveContainer" containerID="81353be09f7749f5b2e22eca0a163e7ec79346eb5cbc71ad52479317f9293c04" Dec 07 09:43:24 crc kubenswrapper[4838]: I1207 09:43:24.749559 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:43:24 crc kubenswrapper[4838]: E1207 09:43:24.749772 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:43:36 crc kubenswrapper[4838]: I1207 09:43:36.614760 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:43:36 crc kubenswrapper[4838]: E1207 09:43:36.615656 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:43:47 crc kubenswrapper[4838]: E1207 09:43:47.080655 4838 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.73:53018->38.102.83.73:33371: write tcp 38.102.83.73:53018->38.102.83.73:33371: write: broken pipe Dec 07 09:43:47 crc kubenswrapper[4838]: I1207 09:43:47.615271 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:43:47 crc kubenswrapper[4838]: E1207 09:43:47.615655 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:44:02 crc kubenswrapper[4838]: I1207 09:44:02.614698 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:44:02 crc kubenswrapper[4838]: E1207 09:44:02.615665 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:44:09 crc kubenswrapper[4838]: I1207 09:44:09.937242 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82"] Dec 07 09:44:09 crc kubenswrapper[4838]: I1207 09:44:09.952871 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jxw82"] Dec 07 09:44:09 crc kubenswrapper[4838]: I1207 09:44:09.970297 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm"] Dec 07 09:44:09 crc kubenswrapper[4838]: I1207 09:44:09.980995 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.017359 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.084877 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.127250 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zqgnn"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.135861 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.145482 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zqgnn"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.150178 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gz247"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.155893 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vdvv4"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.163873 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4jjzm"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.173044 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.184883 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.187597 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k4484"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.195872 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.210038 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-9gr9w"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.215821 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rppbx"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.221053 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-l6q99"] Dec 07 09:44:10 crc kubenswrapper[4838]: I1207 09:44:10.226222 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9l6l6"] Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.626862 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05413639-6dc7-443c-9300-975b60426883" path="/var/lib/kubelet/pods/05413639-6dc7-443c-9300-975b60426883/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.628111 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af361e7-0c07-4c36-95fc-e53fc866b814" path="/var/lib/kubelet/pods/1af361e7-0c07-4c36-95fc-e53fc866b814/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.628756 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b64d9f5-b1db-4ebe-893e-c8da4e2d0297" path="/var/lib/kubelet/pods/1b64d9f5-b1db-4ebe-893e-c8da4e2d0297/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.629469 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bc69631-6ca5-4497-b711-aa0a2d568bbf" path="/var/lib/kubelet/pods/2bc69631-6ca5-4497-b711-aa0a2d568bbf/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.630744 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333d9a84-4e02-4851-b612-0cf2b6677656" path="/var/lib/kubelet/pods/333d9a84-4e02-4851-b612-0cf2b6677656/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.631381 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e" path="/var/lib/kubelet/pods/7b4902c0-f4df-4fd7-aeb8-9e7b172ada4e/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.632001 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e993146-3e36-443e-b88b-1d090aa17900" path="/var/lib/kubelet/pods/9e993146-3e36-443e-b88b-1d090aa17900/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.633193 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c6eb84-9eb9-4101-9de7-dc74da6f4757" path="/var/lib/kubelet/pods/c2c6eb84-9eb9-4101-9de7-dc74da6f4757/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.633934 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2d32b7f-3264-4308-878b-874ccfd18e54" path="/var/lib/kubelet/pods/d2d32b7f-3264-4308-878b-874ccfd18e54/volumes" Dec 07 09:44:11 crc kubenswrapper[4838]: I1207 09:44:11.634588 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e77e2180-8684-4fbd-8b4c-62d07f058ae8" path="/var/lib/kubelet/pods/e77e2180-8684-4fbd-8b4c-62d07f058ae8/volumes" Dec 07 09:44:15 crc kubenswrapper[4838]: I1207 09:44:15.615340 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:44:15 crc kubenswrapper[4838]: E1207 09:44:15.616081 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.180049 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb"] Dec 07 09:44:23 crc kubenswrapper[4838]: E1207 09:44:23.181044 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="registry-server" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181060 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="registry-server" Dec 07 09:44:23 crc kubenswrapper[4838]: E1207 09:44:23.181080 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="registry-server" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181088 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="registry-server" Dec 07 09:44:23 crc kubenswrapper[4838]: E1207 09:44:23.181103 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="extract-content" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181112 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="extract-content" Dec 07 09:44:23 crc kubenswrapper[4838]: E1207 09:44:23.181129 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="extract-utilities" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181136 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="extract-utilities" Dec 07 09:44:23 crc kubenswrapper[4838]: E1207 09:44:23.181152 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="extract-content" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181158 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="extract-content" Dec 07 09:44:23 crc kubenswrapper[4838]: E1207 09:44:23.181174 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="extract-utilities" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181181 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="extract-utilities" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181372 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e1c0746-fd1c-4be4-afe2-6e817cdcb5b5" containerName="registry-server" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.181382 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd723953-667b-4948-948b-d4f1ad03f021" containerName="registry-server" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.182145 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.187990 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.188372 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.188724 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.189067 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.189507 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.208356 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb"] Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.226188 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.226334 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.226366 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kqsf\" (UniqueName: \"kubernetes.io/projected/8587abfa-6182-4613-b458-69c4f1bf2eca-kube-api-access-9kqsf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.226435 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.226467 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.328065 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.328151 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.328192 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kqsf\" (UniqueName: \"kubernetes.io/projected/8587abfa-6182-4613-b458-69c4f1bf2eca-kube-api-access-9kqsf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.328265 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.328296 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.334977 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.335321 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.335845 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.336636 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.359701 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kqsf\" (UniqueName: \"kubernetes.io/projected/8587abfa-6182-4613-b458-69c4f1bf2eca-kube-api-access-9kqsf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:23 crc kubenswrapper[4838]: I1207 09:44:23.508283 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:24 crc kubenswrapper[4838]: W1207 09:44:24.088648 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8587abfa_6182_4613_b458_69c4f1bf2eca.slice/crio-321d3d468c1c60ced39d5117d34665edaa29bd2cf7cff1ef51d783b88fd078a1 WatchSource:0}: Error finding container 321d3d468c1c60ced39d5117d34665edaa29bd2cf7cff1ef51d783b88fd078a1: Status 404 returned error can't find the container with id 321d3d468c1c60ced39d5117d34665edaa29bd2cf7cff1ef51d783b88fd078a1 Dec 07 09:44:24 crc kubenswrapper[4838]: I1207 09:44:24.096345 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb"] Dec 07 09:44:24 crc kubenswrapper[4838]: I1207 09:44:24.295375 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" event={"ID":"8587abfa-6182-4613-b458-69c4f1bf2eca","Type":"ContainerStarted","Data":"321d3d468c1c60ced39d5117d34665edaa29bd2cf7cff1ef51d783b88fd078a1"} Dec 07 09:44:24 crc kubenswrapper[4838]: I1207 09:44:24.694954 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:44:25 crc kubenswrapper[4838]: I1207 09:44:25.313734 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" event={"ID":"8587abfa-6182-4613-b458-69c4f1bf2eca","Type":"ContainerStarted","Data":"2af0d6e6b944b48be51a27d62668883b9c8f557cd566fae4d23b26c439de0275"} Dec 07 09:44:25 crc kubenswrapper[4838]: I1207 09:44:25.337106 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" podStartSLOduration=1.734454178 podStartE2EDuration="2.337081227s" podCreationTimestamp="2025-12-07 09:44:23 +0000 UTC" firstStartedPulling="2025-12-07 09:44:24.090129748 +0000 UTC m=+2280.797448775" lastFinishedPulling="2025-12-07 09:44:24.692756807 +0000 UTC m=+2281.400075824" observedRunningTime="2025-12-07 09:44:25.336051887 +0000 UTC m=+2282.043370964" watchObservedRunningTime="2025-12-07 09:44:25.337081227 +0000 UTC m=+2282.044400284" Dec 07 09:44:27 crc kubenswrapper[4838]: I1207 09:44:27.614458 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:44:27 crc kubenswrapper[4838]: E1207 09:44:27.615283 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:44:37 crc kubenswrapper[4838]: I1207 09:44:37.445352 4838 generic.go:334] "Generic (PLEG): container finished" podID="8587abfa-6182-4613-b458-69c4f1bf2eca" containerID="2af0d6e6b944b48be51a27d62668883b9c8f557cd566fae4d23b26c439de0275" exitCode=0 Dec 07 09:44:37 crc kubenswrapper[4838]: I1207 09:44:37.445442 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" event={"ID":"8587abfa-6182-4613-b458-69c4f1bf2eca","Type":"ContainerDied","Data":"2af0d6e6b944b48be51a27d62668883b9c8f557cd566fae4d23b26c439de0275"} Dec 07 09:44:38 crc kubenswrapper[4838]: I1207 09:44:38.923562 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.057951 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ceph\") pod \"8587abfa-6182-4613-b458-69c4f1bf2eca\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.058103 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kqsf\" (UniqueName: \"kubernetes.io/projected/8587abfa-6182-4613-b458-69c4f1bf2eca-kube-api-access-9kqsf\") pod \"8587abfa-6182-4613-b458-69c4f1bf2eca\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.058125 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ssh-key\") pod \"8587abfa-6182-4613-b458-69c4f1bf2eca\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.058196 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-repo-setup-combined-ca-bundle\") pod \"8587abfa-6182-4613-b458-69c4f1bf2eca\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.058289 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-inventory\") pod \"8587abfa-6182-4613-b458-69c4f1bf2eca\" (UID: \"8587abfa-6182-4613-b458-69c4f1bf2eca\") " Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.063950 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ceph" (OuterVolumeSpecName: "ceph") pod "8587abfa-6182-4613-b458-69c4f1bf2eca" (UID: "8587abfa-6182-4613-b458-69c4f1bf2eca"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.065305 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8587abfa-6182-4613-b458-69c4f1bf2eca-kube-api-access-9kqsf" (OuterVolumeSpecName: "kube-api-access-9kqsf") pod "8587abfa-6182-4613-b458-69c4f1bf2eca" (UID: "8587abfa-6182-4613-b458-69c4f1bf2eca"). InnerVolumeSpecName "kube-api-access-9kqsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.080797 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8587abfa-6182-4613-b458-69c4f1bf2eca" (UID: "8587abfa-6182-4613-b458-69c4f1bf2eca"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.085777 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-inventory" (OuterVolumeSpecName: "inventory") pod "8587abfa-6182-4613-b458-69c4f1bf2eca" (UID: "8587abfa-6182-4613-b458-69c4f1bf2eca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.089988 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8587abfa-6182-4613-b458-69c4f1bf2eca" (UID: "8587abfa-6182-4613-b458-69c4f1bf2eca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.161087 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kqsf\" (UniqueName: \"kubernetes.io/projected/8587abfa-6182-4613-b458-69c4f1bf2eca-kube-api-access-9kqsf\") on node \"crc\" DevicePath \"\"" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.161120 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.161132 4838 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.161145 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.161158 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8587abfa-6182-4613-b458-69c4f1bf2eca-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.473227 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" event={"ID":"8587abfa-6182-4613-b458-69c4f1bf2eca","Type":"ContainerDied","Data":"321d3d468c1c60ced39d5117d34665edaa29bd2cf7cff1ef51d783b88fd078a1"} Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.473756 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="321d3d468c1c60ced39d5117d34665edaa29bd2cf7cff1ef51d783b88fd078a1" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.473883 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.560590 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd"] Dec 07 09:44:39 crc kubenswrapper[4838]: E1207 09:44:39.561011 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8587abfa-6182-4613-b458-69c4f1bf2eca" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.561031 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8587abfa-6182-4613-b458-69c4f1bf2eca" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.561245 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8587abfa-6182-4613-b458-69c4f1bf2eca" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.561996 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.564491 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.564793 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.565101 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.565245 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.570392 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.575180 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd"] Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.614576 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:44:39 crc kubenswrapper[4838]: E1207 09:44:39.614835 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.673002 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj6t9\" (UniqueName: \"kubernetes.io/projected/1386eb32-0cef-4287-93d1-8fb5445081f3-kube-api-access-zj6t9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.673049 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.673068 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.673095 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.673125 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.774715 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj6t9\" (UniqueName: \"kubernetes.io/projected/1386eb32-0cef-4287-93d1-8fb5445081f3-kube-api-access-zj6t9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.774763 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.774792 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.774851 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.774886 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.779899 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.780513 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.781312 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.782193 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.791487 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj6t9\" (UniqueName: \"kubernetes.io/projected/1386eb32-0cef-4287-93d1-8fb5445081f3-kube-api-access-zj6t9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:39 crc kubenswrapper[4838]: I1207 09:44:39.892798 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:44:40 crc kubenswrapper[4838]: W1207 09:44:40.418484 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1386eb32_0cef_4287_93d1_8fb5445081f3.slice/crio-5832cbbad4ccd40630d4dc8ad90c430d389abd3b49480ee73346fe5e2a181bef WatchSource:0}: Error finding container 5832cbbad4ccd40630d4dc8ad90c430d389abd3b49480ee73346fe5e2a181bef: Status 404 returned error can't find the container with id 5832cbbad4ccd40630d4dc8ad90c430d389abd3b49480ee73346fe5e2a181bef Dec 07 09:44:40 crc kubenswrapper[4838]: I1207 09:44:40.421425 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd"] Dec 07 09:44:40 crc kubenswrapper[4838]: I1207 09:44:40.480405 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" event={"ID":"1386eb32-0cef-4287-93d1-8fb5445081f3","Type":"ContainerStarted","Data":"5832cbbad4ccd40630d4dc8ad90c430d389abd3b49480ee73346fe5e2a181bef"} Dec 07 09:44:41 crc kubenswrapper[4838]: I1207 09:44:41.494864 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" event={"ID":"1386eb32-0cef-4287-93d1-8fb5445081f3","Type":"ContainerStarted","Data":"cbca5c3298976a80cc5f7b588b69da14e5f1cd0fec6d842081ae8a1376d82c6c"} Dec 07 09:44:41 crc kubenswrapper[4838]: I1207 09:44:41.527505 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" podStartSLOduration=2.124352909 podStartE2EDuration="2.527479192s" podCreationTimestamp="2025-12-07 09:44:39 +0000 UTC" firstStartedPulling="2025-12-07 09:44:40.420317798 +0000 UTC m=+2297.127636815" lastFinishedPulling="2025-12-07 09:44:40.823444081 +0000 UTC m=+2297.530763098" observedRunningTime="2025-12-07 09:44:41.515516372 +0000 UTC m=+2298.222835409" watchObservedRunningTime="2025-12-07 09:44:41.527479192 +0000 UTC m=+2298.234798239" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.035019 4838 scope.go:117] "RemoveContainer" containerID="df7a2b3f98440574d1a9abf0b1655c31d3286fdeec229e3e5a4fa10be3ed191b" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.131127 4838 scope.go:117] "RemoveContainer" containerID="19d85c270699df25aee8183fa0de7b595a3af3a8a51966a213675c01c342beb3" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.266215 4838 scope.go:117] "RemoveContainer" containerID="b6073d5b04ed3abcb072c8855b1b69665acbf3072734d569e28ac30635a5d783" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.332491 4838 scope.go:117] "RemoveContainer" containerID="5ed677d43f8bccdeaafc520051c7efccfc880e4a12cd32f48bf2014cd801e02a" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.362783 4838 scope.go:117] "RemoveContainer" containerID="a5a982b3242fa6920d9f10ca35ad81cc37c5c4395a66cf9a1cc73ffbd87e19d1" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.414313 4838 scope.go:117] "RemoveContainer" containerID="5eb0970e21d534da8fa970ee8241ea8e67ae26d54d9e78aa3b69595227fd1e55" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.442960 4838 scope.go:117] "RemoveContainer" containerID="967cd965ac848b511b7edf75c70dcfdb50b08a6b2bbab12248ae3712810291c8" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.467395 4838 scope.go:117] "RemoveContainer" containerID="d1ce0601d6eaf316219ec31f5880f8ee80f783397491ed2653377f89677e5455" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.521392 4838 scope.go:117] "RemoveContainer" containerID="faba245eb60d21eacae0f9eaef532b5e0d4df4c6bdde2d35404659cbfd4ee0c4" Dec 07 09:44:52 crc kubenswrapper[4838]: I1207 09:44:52.555091 4838 scope.go:117] "RemoveContainer" containerID="cf921b0369ddf99c99cede84e0acf6c60c2005f8725e2b2af97ba7f5d8465570" Dec 07 09:44:54 crc kubenswrapper[4838]: I1207 09:44:54.615129 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:44:54 crc kubenswrapper[4838]: E1207 09:44:54.615585 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.157849 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg"] Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.159770 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.164982 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.167046 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.181516 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg"] Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.293344 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aadda5c8-2133-474c-a880-4a3c519bf748-secret-volume\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.293744 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xg7v\" (UniqueName: \"kubernetes.io/projected/aadda5c8-2133-474c-a880-4a3c519bf748-kube-api-access-7xg7v\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.294181 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aadda5c8-2133-474c-a880-4a3c519bf748-config-volume\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.395941 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aadda5c8-2133-474c-a880-4a3c519bf748-config-volume\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.395999 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aadda5c8-2133-474c-a880-4a3c519bf748-secret-volume\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.396101 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xg7v\" (UniqueName: \"kubernetes.io/projected/aadda5c8-2133-474c-a880-4a3c519bf748-kube-api-access-7xg7v\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.397535 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aadda5c8-2133-474c-a880-4a3c519bf748-config-volume\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.412052 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aadda5c8-2133-474c-a880-4a3c519bf748-secret-volume\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.414716 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xg7v\" (UniqueName: \"kubernetes.io/projected/aadda5c8-2133-474c-a880-4a3c519bf748-kube-api-access-7xg7v\") pod \"collect-profiles-29418345-m6chg\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.532527 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:00 crc kubenswrapper[4838]: I1207 09:45:00.980460 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg"] Dec 07 09:45:01 crc kubenswrapper[4838]: I1207 09:45:01.738385 4838 generic.go:334] "Generic (PLEG): container finished" podID="aadda5c8-2133-474c-a880-4a3c519bf748" containerID="3cadfad733aeffaa303627fe927a40291abae7f178500385ddfaf21783466b54" exitCode=0 Dec 07 09:45:01 crc kubenswrapper[4838]: I1207 09:45:01.738470 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" event={"ID":"aadda5c8-2133-474c-a880-4a3c519bf748","Type":"ContainerDied","Data":"3cadfad733aeffaa303627fe927a40291abae7f178500385ddfaf21783466b54"} Dec 07 09:45:01 crc kubenswrapper[4838]: I1207 09:45:01.738796 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" event={"ID":"aadda5c8-2133-474c-a880-4a3c519bf748","Type":"ContainerStarted","Data":"00a03823b07bb313be61898d7657974b80002e8c11fa4ea01383f1a7b70e8e88"} Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.105046 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.149691 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aadda5c8-2133-474c-a880-4a3c519bf748-config-volume\") pod \"aadda5c8-2133-474c-a880-4a3c519bf748\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.149855 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xg7v\" (UniqueName: \"kubernetes.io/projected/aadda5c8-2133-474c-a880-4a3c519bf748-kube-api-access-7xg7v\") pod \"aadda5c8-2133-474c-a880-4a3c519bf748\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.149898 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aadda5c8-2133-474c-a880-4a3c519bf748-secret-volume\") pod \"aadda5c8-2133-474c-a880-4a3c519bf748\" (UID: \"aadda5c8-2133-474c-a880-4a3c519bf748\") " Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.151260 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aadda5c8-2133-474c-a880-4a3c519bf748-config-volume" (OuterVolumeSpecName: "config-volume") pod "aadda5c8-2133-474c-a880-4a3c519bf748" (UID: "aadda5c8-2133-474c-a880-4a3c519bf748"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.155979 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aadda5c8-2133-474c-a880-4a3c519bf748-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aadda5c8-2133-474c-a880-4a3c519bf748" (UID: "aadda5c8-2133-474c-a880-4a3c519bf748"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.156574 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aadda5c8-2133-474c-a880-4a3c519bf748-kube-api-access-7xg7v" (OuterVolumeSpecName: "kube-api-access-7xg7v") pod "aadda5c8-2133-474c-a880-4a3c519bf748" (UID: "aadda5c8-2133-474c-a880-4a3c519bf748"). InnerVolumeSpecName "kube-api-access-7xg7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.252225 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aadda5c8-2133-474c-a880-4a3c519bf748-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.252265 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xg7v\" (UniqueName: \"kubernetes.io/projected/aadda5c8-2133-474c-a880-4a3c519bf748-kube-api-access-7xg7v\") on node \"crc\" DevicePath \"\"" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.252280 4838 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aadda5c8-2133-474c-a880-4a3c519bf748-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.759132 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" event={"ID":"aadda5c8-2133-474c-a880-4a3c519bf748","Type":"ContainerDied","Data":"00a03823b07bb313be61898d7657974b80002e8c11fa4ea01383f1a7b70e8e88"} Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.759181 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00a03823b07bb313be61898d7657974b80002e8c11fa4ea01383f1a7b70e8e88" Dec 07 09:45:03 crc kubenswrapper[4838]: I1207 09:45:03.759184 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg" Dec 07 09:45:04 crc kubenswrapper[4838]: I1207 09:45:04.192941 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm"] Dec 07 09:45:04 crc kubenswrapper[4838]: I1207 09:45:04.200992 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418300-jr5mm"] Dec 07 09:45:05 crc kubenswrapper[4838]: I1207 09:45:05.626589 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11c62d66-9fdc-480e-b259-dc768c6a04e5" path="/var/lib/kubelet/pods/11c62d66-9fdc-480e-b259-dc768c6a04e5/volumes" Dec 07 09:45:06 crc kubenswrapper[4838]: I1207 09:45:06.615019 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:45:06 crc kubenswrapper[4838]: E1207 09:45:06.615323 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:45:17 crc kubenswrapper[4838]: I1207 09:45:17.615348 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:45:17 crc kubenswrapper[4838]: E1207 09:45:17.616564 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:45:31 crc kubenswrapper[4838]: I1207 09:45:31.614259 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:45:31 crc kubenswrapper[4838]: E1207 09:45:31.615070 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:45:42 crc kubenswrapper[4838]: I1207 09:45:42.615187 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:45:42 crc kubenswrapper[4838]: E1207 09:45:42.615934 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:45:52 crc kubenswrapper[4838]: I1207 09:45:52.725906 4838 scope.go:117] "RemoveContainer" containerID="580f2f9da5cc086f3c110e37b68671eac79b7c85210619f74a10d3060a0f25eb" Dec 07 09:45:57 crc kubenswrapper[4838]: I1207 09:45:57.615128 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:45:57 crc kubenswrapper[4838]: E1207 09:45:57.615922 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:46:11 crc kubenswrapper[4838]: I1207 09:46:11.615942 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:46:11 crc kubenswrapper[4838]: E1207 09:46:11.616890 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:46:26 crc kubenswrapper[4838]: I1207 09:46:26.615496 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:46:26 crc kubenswrapper[4838]: E1207 09:46:26.616279 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:46:38 crc kubenswrapper[4838]: I1207 09:46:38.615318 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:46:38 crc kubenswrapper[4838]: E1207 09:46:38.616681 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:46:46 crc kubenswrapper[4838]: I1207 09:46:46.669650 4838 generic.go:334] "Generic (PLEG): container finished" podID="1386eb32-0cef-4287-93d1-8fb5445081f3" containerID="cbca5c3298976a80cc5f7b588b69da14e5f1cd0fec6d842081ae8a1376d82c6c" exitCode=0 Dec 07 09:46:46 crc kubenswrapper[4838]: I1207 09:46:46.669732 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" event={"ID":"1386eb32-0cef-4287-93d1-8fb5445081f3","Type":"ContainerDied","Data":"cbca5c3298976a80cc5f7b588b69da14e5f1cd0fec6d842081ae8a1376d82c6c"} Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.125549 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.273997 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-bootstrap-combined-ca-bundle\") pod \"1386eb32-0cef-4287-93d1-8fb5445081f3\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.274150 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-inventory\") pod \"1386eb32-0cef-4287-93d1-8fb5445081f3\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.274265 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ssh-key\") pod \"1386eb32-0cef-4287-93d1-8fb5445081f3\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.274428 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ceph\") pod \"1386eb32-0cef-4287-93d1-8fb5445081f3\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.274485 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj6t9\" (UniqueName: \"kubernetes.io/projected/1386eb32-0cef-4287-93d1-8fb5445081f3-kube-api-access-zj6t9\") pod \"1386eb32-0cef-4287-93d1-8fb5445081f3\" (UID: \"1386eb32-0cef-4287-93d1-8fb5445081f3\") " Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.287714 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386eb32-0cef-4287-93d1-8fb5445081f3-kube-api-access-zj6t9" (OuterVolumeSpecName: "kube-api-access-zj6t9") pod "1386eb32-0cef-4287-93d1-8fb5445081f3" (UID: "1386eb32-0cef-4287-93d1-8fb5445081f3"). InnerVolumeSpecName "kube-api-access-zj6t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.290917 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ceph" (OuterVolumeSpecName: "ceph") pod "1386eb32-0cef-4287-93d1-8fb5445081f3" (UID: "1386eb32-0cef-4287-93d1-8fb5445081f3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.298917 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1386eb32-0cef-4287-93d1-8fb5445081f3" (UID: "1386eb32-0cef-4287-93d1-8fb5445081f3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.311920 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1386eb32-0cef-4287-93d1-8fb5445081f3" (UID: "1386eb32-0cef-4287-93d1-8fb5445081f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.335938 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-inventory" (OuterVolumeSpecName: "inventory") pod "1386eb32-0cef-4287-93d1-8fb5445081f3" (UID: "1386eb32-0cef-4287-93d1-8fb5445081f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.376476 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.376715 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.376798 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj6t9\" (UniqueName: \"kubernetes.io/projected/1386eb32-0cef-4287-93d1-8fb5445081f3-kube-api-access-zj6t9\") on node \"crc\" DevicePath \"\"" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.376899 4838 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.376988 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1386eb32-0cef-4287-93d1-8fb5445081f3-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.688874 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" event={"ID":"1386eb32-0cef-4287-93d1-8fb5445081f3","Type":"ContainerDied","Data":"5832cbbad4ccd40630d4dc8ad90c430d389abd3b49480ee73346fe5e2a181bef"} Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.688963 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5832cbbad4ccd40630d4dc8ad90c430d389abd3b49480ee73346fe5e2a181bef" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.689015 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.793670 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk"] Dec 07 09:46:48 crc kubenswrapper[4838]: E1207 09:46:48.794064 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1386eb32-0cef-4287-93d1-8fb5445081f3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.794084 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="1386eb32-0cef-4287-93d1-8fb5445081f3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 09:46:48 crc kubenswrapper[4838]: E1207 09:46:48.794097 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aadda5c8-2133-474c-a880-4a3c519bf748" containerName="collect-profiles" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.794104 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="aadda5c8-2133-474c-a880-4a3c519bf748" containerName="collect-profiles" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.794273 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="aadda5c8-2133-474c-a880-4a3c519bf748" containerName="collect-profiles" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.794299 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="1386eb32-0cef-4287-93d1-8fb5445081f3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.794857 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.797122 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.797347 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.797755 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.798584 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.808212 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk"] Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.808772 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.987100 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.987141 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k964m\" (UniqueName: \"kubernetes.io/projected/f5b4b264-d902-41a3-a100-6455a8a7b6ca-kube-api-access-k964m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.987255 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:48 crc kubenswrapper[4838]: I1207 09:46:48.987294 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.089004 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.089744 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k964m\" (UniqueName: \"kubernetes.io/projected/f5b4b264-d902-41a3-a100-6455a8a7b6ca-kube-api-access-k964m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.090020 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.090179 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.097057 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.107928 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.110592 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.111157 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k964m\" (UniqueName: \"kubernetes.io/projected/f5b4b264-d902-41a3-a100-6455a8a7b6ca-kube-api-access-k964m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.116530 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.700220 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk"] Dec 07 09:46:49 crc kubenswrapper[4838]: I1207 09:46:49.718659 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:46:50 crc kubenswrapper[4838]: I1207 09:46:50.614476 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:46:50 crc kubenswrapper[4838]: E1207 09:46:50.615129 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:46:50 crc kubenswrapper[4838]: I1207 09:46:50.707805 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" event={"ID":"f5b4b264-d902-41a3-a100-6455a8a7b6ca","Type":"ContainerStarted","Data":"d85a7b87d373a993ba30bef26417c0861951cd41f167233624720c10161e201b"} Dec 07 09:46:50 crc kubenswrapper[4838]: I1207 09:46:50.710385 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" event={"ID":"f5b4b264-d902-41a3-a100-6455a8a7b6ca","Type":"ContainerStarted","Data":"ff923608864ec0e093bb69c132095ba94f9d8c2021867e7d828f51f1b8878d9d"} Dec 07 09:46:50 crc kubenswrapper[4838]: I1207 09:46:50.730322 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" podStartSLOduration=2.216128974 podStartE2EDuration="2.730302365s" podCreationTimestamp="2025-12-07 09:46:48 +0000 UTC" firstStartedPulling="2025-12-07 09:46:49.718421045 +0000 UTC m=+2426.425740062" lastFinishedPulling="2025-12-07 09:46:50.232594436 +0000 UTC m=+2426.939913453" observedRunningTime="2025-12-07 09:46:50.722420195 +0000 UTC m=+2427.429739212" watchObservedRunningTime="2025-12-07 09:46:50.730302365 +0000 UTC m=+2427.437621382" Dec 07 09:47:02 crc kubenswrapper[4838]: I1207 09:47:02.614487 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:47:02 crc kubenswrapper[4838]: E1207 09:47:02.615185 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:47:14 crc kubenswrapper[4838]: I1207 09:47:14.615697 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:47:14 crc kubenswrapper[4838]: E1207 09:47:14.616581 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:47:21 crc kubenswrapper[4838]: I1207 09:47:21.001805 4838 generic.go:334] "Generic (PLEG): container finished" podID="f5b4b264-d902-41a3-a100-6455a8a7b6ca" containerID="d85a7b87d373a993ba30bef26417c0861951cd41f167233624720c10161e201b" exitCode=0 Dec 07 09:47:21 crc kubenswrapper[4838]: I1207 09:47:21.001954 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" event={"ID":"f5b4b264-d902-41a3-a100-6455a8a7b6ca","Type":"ContainerDied","Data":"d85a7b87d373a993ba30bef26417c0861951cd41f167233624720c10161e201b"} Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.462694 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.541723 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k964m\" (UniqueName: \"kubernetes.io/projected/f5b4b264-d902-41a3-a100-6455a8a7b6ca-kube-api-access-k964m\") pod \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.541785 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ssh-key\") pod \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.541911 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ceph\") pod \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.541952 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-inventory\") pod \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\" (UID: \"f5b4b264-d902-41a3-a100-6455a8a7b6ca\") " Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.547969 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ceph" (OuterVolumeSpecName: "ceph") pod "f5b4b264-d902-41a3-a100-6455a8a7b6ca" (UID: "f5b4b264-d902-41a3-a100-6455a8a7b6ca"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.554964 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b4b264-d902-41a3-a100-6455a8a7b6ca-kube-api-access-k964m" (OuterVolumeSpecName: "kube-api-access-k964m") pod "f5b4b264-d902-41a3-a100-6455a8a7b6ca" (UID: "f5b4b264-d902-41a3-a100-6455a8a7b6ca"). InnerVolumeSpecName "kube-api-access-k964m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.573678 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f5b4b264-d902-41a3-a100-6455a8a7b6ca" (UID: "f5b4b264-d902-41a3-a100-6455a8a7b6ca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.576447 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-inventory" (OuterVolumeSpecName: "inventory") pod "f5b4b264-d902-41a3-a100-6455a8a7b6ca" (UID: "f5b4b264-d902-41a3-a100-6455a8a7b6ca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.643200 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.643231 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.643244 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k964m\" (UniqueName: \"kubernetes.io/projected/f5b4b264-d902-41a3-a100-6455a8a7b6ca-kube-api-access-k964m\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:22 crc kubenswrapper[4838]: I1207 09:47:22.643255 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5b4b264-d902-41a3-a100-6455a8a7b6ca-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.020612 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" event={"ID":"f5b4b264-d902-41a3-a100-6455a8a7b6ca","Type":"ContainerDied","Data":"ff923608864ec0e093bb69c132095ba94f9d8c2021867e7d828f51f1b8878d9d"} Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.020929 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff923608864ec0e093bb69c132095ba94f9d8c2021867e7d828f51f1b8878d9d" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.020674 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.130403 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk"] Dec 07 09:47:23 crc kubenswrapper[4838]: E1207 09:47:23.130879 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b4b264-d902-41a3-a100-6455a8a7b6ca" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.130906 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b4b264-d902-41a3-a100-6455a8a7b6ca" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.131203 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5b4b264-d902-41a3-a100-6455a8a7b6ca" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.133574 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.138868 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.140206 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.140421 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.140535 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.140838 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.149037 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk"] Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.151633 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt9z7\" (UniqueName: \"kubernetes.io/projected/f9c17708-e8da-45b1-b919-354138adf40b-kube-api-access-gt9z7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.151726 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.151958 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.152025 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.253803 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.253922 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.253954 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.254028 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt9z7\" (UniqueName: \"kubernetes.io/projected/f9c17708-e8da-45b1-b919-354138adf40b-kube-api-access-gt9z7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.258665 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.258683 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.265105 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.273015 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt9z7\" (UniqueName: \"kubernetes.io/projected/f9c17708-e8da-45b1-b919-354138adf40b-kube-api-access-gt9z7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-99vbk\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:23 crc kubenswrapper[4838]: I1207 09:47:23.456206 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:24 crc kubenswrapper[4838]: I1207 09:47:24.052569 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk"] Dec 07 09:47:24 crc kubenswrapper[4838]: I1207 09:47:24.515190 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:47:25 crc kubenswrapper[4838]: I1207 09:47:25.047958 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" event={"ID":"f9c17708-e8da-45b1-b919-354138adf40b","Type":"ContainerStarted","Data":"0d8b808e957b719d59156e20ed1a3f032ee59d3c3548be60f096a437a1736082"} Dec 07 09:47:25 crc kubenswrapper[4838]: I1207 09:47:25.048350 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" event={"ID":"f9c17708-e8da-45b1-b919-354138adf40b","Type":"ContainerStarted","Data":"f5a1604589c309d6454672529c78ef1f8515466ef59f9517c6a1cdc06a4c192a"} Dec 07 09:47:25 crc kubenswrapper[4838]: I1207 09:47:25.072290 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" podStartSLOduration=1.614894353 podStartE2EDuration="2.07226882s" podCreationTimestamp="2025-12-07 09:47:23 +0000 UTC" firstStartedPulling="2025-12-07 09:47:24.053939606 +0000 UTC m=+2460.761258623" lastFinishedPulling="2025-12-07 09:47:24.511314033 +0000 UTC m=+2461.218633090" observedRunningTime="2025-12-07 09:47:25.069357405 +0000 UTC m=+2461.776676442" watchObservedRunningTime="2025-12-07 09:47:25.07226882 +0000 UTC m=+2461.779587857" Dec 07 09:47:28 crc kubenswrapper[4838]: I1207 09:47:28.614423 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:47:28 crc kubenswrapper[4838]: E1207 09:47:28.615095 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:47:30 crc kubenswrapper[4838]: I1207 09:47:30.097593 4838 generic.go:334] "Generic (PLEG): container finished" podID="f9c17708-e8da-45b1-b919-354138adf40b" containerID="0d8b808e957b719d59156e20ed1a3f032ee59d3c3548be60f096a437a1736082" exitCode=0 Dec 07 09:47:30 crc kubenswrapper[4838]: I1207 09:47:30.097681 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" event={"ID":"f9c17708-e8da-45b1-b919-354138adf40b","Type":"ContainerDied","Data":"0d8b808e957b719d59156e20ed1a3f032ee59d3c3548be60f096a437a1736082"} Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.481002 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.534677 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ssh-key\") pod \"f9c17708-e8da-45b1-b919-354138adf40b\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.535177 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt9z7\" (UniqueName: \"kubernetes.io/projected/f9c17708-e8da-45b1-b919-354138adf40b-kube-api-access-gt9z7\") pod \"f9c17708-e8da-45b1-b919-354138adf40b\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.535307 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-inventory\") pod \"f9c17708-e8da-45b1-b919-354138adf40b\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.535467 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ceph\") pod \"f9c17708-e8da-45b1-b919-354138adf40b\" (UID: \"f9c17708-e8da-45b1-b919-354138adf40b\") " Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.540170 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9c17708-e8da-45b1-b919-354138adf40b-kube-api-access-gt9z7" (OuterVolumeSpecName: "kube-api-access-gt9z7") pod "f9c17708-e8da-45b1-b919-354138adf40b" (UID: "f9c17708-e8da-45b1-b919-354138adf40b"). InnerVolumeSpecName "kube-api-access-gt9z7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.540651 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ceph" (OuterVolumeSpecName: "ceph") pod "f9c17708-e8da-45b1-b919-354138adf40b" (UID: "f9c17708-e8da-45b1-b919-354138adf40b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.563010 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9c17708-e8da-45b1-b919-354138adf40b" (UID: "f9c17708-e8da-45b1-b919-354138adf40b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.564222 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-inventory" (OuterVolumeSpecName: "inventory") pod "f9c17708-e8da-45b1-b919-354138adf40b" (UID: "f9c17708-e8da-45b1-b919-354138adf40b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.637454 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.637768 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt9z7\" (UniqueName: \"kubernetes.io/projected/f9c17708-e8da-45b1-b919-354138adf40b-kube-api-access-gt9z7\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.637778 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:31 crc kubenswrapper[4838]: I1207 09:47:31.637786 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9c17708-e8da-45b1-b919-354138adf40b-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.120374 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" event={"ID":"f9c17708-e8da-45b1-b919-354138adf40b","Type":"ContainerDied","Data":"f5a1604589c309d6454672529c78ef1f8515466ef59f9517c6a1cdc06a4c192a"} Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.120461 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5a1604589c309d6454672529c78ef1f8515466ef59f9517c6a1cdc06a4c192a" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.120585 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-99vbk" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.193463 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt"] Dec 07 09:47:32 crc kubenswrapper[4838]: E1207 09:47:32.193856 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c17708-e8da-45b1-b919-354138adf40b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.193872 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c17708-e8da-45b1-b919-354138adf40b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.194033 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c17708-e8da-45b1-b919-354138adf40b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.194681 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.198836 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.199191 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.199518 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.199672 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.214098 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.217053 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt"] Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.251042 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgpcm\" (UniqueName: \"kubernetes.io/projected/6b76e0a9-db0a-4913-90a7-1306215b82b7-kube-api-access-tgpcm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.251146 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.251231 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.251265 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.353033 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.353087 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.353143 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgpcm\" (UniqueName: \"kubernetes.io/projected/6b76e0a9-db0a-4913-90a7-1306215b82b7-kube-api-access-tgpcm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.353205 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.357789 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.362309 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.366284 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.371205 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgpcm\" (UniqueName: \"kubernetes.io/projected/6b76e0a9-db0a-4913-90a7-1306215b82b7-kube-api-access-tgpcm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nlmxt\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:32 crc kubenswrapper[4838]: I1207 09:47:32.520785 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:47:33 crc kubenswrapper[4838]: I1207 09:47:33.020315 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt"] Dec 07 09:47:33 crc kubenswrapper[4838]: I1207 09:47:33.134148 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" event={"ID":"6b76e0a9-db0a-4913-90a7-1306215b82b7","Type":"ContainerStarted","Data":"2698173b0ef3fcf739a5df468d8cdcaa8264646bd47c7753dea7ee42bb19c0cf"} Dec 07 09:47:34 crc kubenswrapper[4838]: I1207 09:47:34.146768 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" event={"ID":"6b76e0a9-db0a-4913-90a7-1306215b82b7","Type":"ContainerStarted","Data":"e3faac10512d30fd998b4f07de98efd8350ba86dbbd9eace8c60d9c81325dba8"} Dec 07 09:47:34 crc kubenswrapper[4838]: I1207 09:47:34.168308 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" podStartSLOduration=1.506580759 podStartE2EDuration="2.168289953s" podCreationTimestamp="2025-12-07 09:47:32 +0000 UTC" firstStartedPulling="2025-12-07 09:47:33.02819544 +0000 UTC m=+2469.735514467" lastFinishedPulling="2025-12-07 09:47:33.689904644 +0000 UTC m=+2470.397223661" observedRunningTime="2025-12-07 09:47:34.163104162 +0000 UTC m=+2470.870423189" watchObservedRunningTime="2025-12-07 09:47:34.168289953 +0000 UTC m=+2470.875608970" Dec 07 09:47:39 crc kubenswrapper[4838]: I1207 09:47:39.614888 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:47:39 crc kubenswrapper[4838]: E1207 09:47:39.615585 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:47:54 crc kubenswrapper[4838]: I1207 09:47:54.615354 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:47:54 crc kubenswrapper[4838]: E1207 09:47:54.616684 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:48:09 crc kubenswrapper[4838]: I1207 09:48:09.615254 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:48:09 crc kubenswrapper[4838]: E1207 09:48:09.616369 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:48:20 crc kubenswrapper[4838]: I1207 09:48:20.516231 4838 generic.go:334] "Generic (PLEG): container finished" podID="6b76e0a9-db0a-4913-90a7-1306215b82b7" containerID="e3faac10512d30fd998b4f07de98efd8350ba86dbbd9eace8c60d9c81325dba8" exitCode=0 Dec 07 09:48:20 crc kubenswrapper[4838]: I1207 09:48:20.516323 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" event={"ID":"6b76e0a9-db0a-4913-90a7-1306215b82b7","Type":"ContainerDied","Data":"e3faac10512d30fd998b4f07de98efd8350ba86dbbd9eace8c60d9c81325dba8"} Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.020513 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.148031 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ssh-key\") pod \"6b76e0a9-db0a-4913-90a7-1306215b82b7\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.148232 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-inventory\") pod \"6b76e0a9-db0a-4913-90a7-1306215b82b7\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.148323 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ceph\") pod \"6b76e0a9-db0a-4913-90a7-1306215b82b7\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.148378 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgpcm\" (UniqueName: \"kubernetes.io/projected/6b76e0a9-db0a-4913-90a7-1306215b82b7-kube-api-access-tgpcm\") pod \"6b76e0a9-db0a-4913-90a7-1306215b82b7\" (UID: \"6b76e0a9-db0a-4913-90a7-1306215b82b7\") " Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.153505 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ceph" (OuterVolumeSpecName: "ceph") pod "6b76e0a9-db0a-4913-90a7-1306215b82b7" (UID: "6b76e0a9-db0a-4913-90a7-1306215b82b7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.154151 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b76e0a9-db0a-4913-90a7-1306215b82b7-kube-api-access-tgpcm" (OuterVolumeSpecName: "kube-api-access-tgpcm") pod "6b76e0a9-db0a-4913-90a7-1306215b82b7" (UID: "6b76e0a9-db0a-4913-90a7-1306215b82b7"). InnerVolumeSpecName "kube-api-access-tgpcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.176315 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-inventory" (OuterVolumeSpecName: "inventory") pod "6b76e0a9-db0a-4913-90a7-1306215b82b7" (UID: "6b76e0a9-db0a-4913-90a7-1306215b82b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.198968 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b76e0a9-db0a-4913-90a7-1306215b82b7" (UID: "6b76e0a9-db0a-4913-90a7-1306215b82b7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.250516 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.250557 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.250571 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b76e0a9-db0a-4913-90a7-1306215b82b7-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.250584 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgpcm\" (UniqueName: \"kubernetes.io/projected/6b76e0a9-db0a-4913-90a7-1306215b82b7-kube-api-access-tgpcm\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.555749 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" event={"ID":"6b76e0a9-db0a-4913-90a7-1306215b82b7","Type":"ContainerDied","Data":"2698173b0ef3fcf739a5df468d8cdcaa8264646bd47c7753dea7ee42bb19c0cf"} Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.555787 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2698173b0ef3fcf739a5df468d8cdcaa8264646bd47c7753dea7ee42bb19c0cf" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.555855 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nlmxt" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.651747 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56"] Dec 07 09:48:22 crc kubenswrapper[4838]: E1207 09:48:22.652180 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b76e0a9-db0a-4913-90a7-1306215b82b7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.652197 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b76e0a9-db0a-4913-90a7-1306215b82b7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.652359 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b76e0a9-db0a-4913-90a7-1306215b82b7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.652957 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.657169 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.657338 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.657346 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.657510 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.659671 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.666974 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56"] Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.759347 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.759439 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.760133 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7676d\" (UniqueName: \"kubernetes.io/projected/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-kube-api-access-7676d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.760288 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.862029 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7676d\" (UniqueName: \"kubernetes.io/projected/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-kube-api-access-7676d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.862122 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.862163 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.862205 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.867440 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.868944 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.869716 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.889900 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7676d\" (UniqueName: \"kubernetes.io/projected/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-kube-api-access-7676d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:22 crc kubenswrapper[4838]: I1207 09:48:22.981310 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:23 crc kubenswrapper[4838]: I1207 09:48:23.497661 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56"] Dec 07 09:48:23 crc kubenswrapper[4838]: I1207 09:48:23.563733 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" event={"ID":"9e6c35ea-c607-4c61-84d0-7bdf0d62add9","Type":"ContainerStarted","Data":"4f87982f41ed74f6b44024f17e546e8f97e1d6bed1b20095b23082afa0030e9f"} Dec 07 09:48:24 crc kubenswrapper[4838]: I1207 09:48:24.267771 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:48:24 crc kubenswrapper[4838]: I1207 09:48:24.571623 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" event={"ID":"9e6c35ea-c607-4c61-84d0-7bdf0d62add9","Type":"ContainerStarted","Data":"74d554e9726f99037638d695efda0e16eeb9beb22388e39db7871beca13265c6"} Dec 07 09:48:24 crc kubenswrapper[4838]: I1207 09:48:24.592835 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" podStartSLOduration=1.851139049 podStartE2EDuration="2.592793156s" podCreationTimestamp="2025-12-07 09:48:22 +0000 UTC" firstStartedPulling="2025-12-07 09:48:23.52348333 +0000 UTC m=+2520.230802357" lastFinishedPulling="2025-12-07 09:48:24.265137437 +0000 UTC m=+2520.972456464" observedRunningTime="2025-12-07 09:48:24.591954102 +0000 UTC m=+2521.299273119" watchObservedRunningTime="2025-12-07 09:48:24.592793156 +0000 UTC m=+2521.300112173" Dec 07 09:48:24 crc kubenswrapper[4838]: I1207 09:48:24.614830 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:48:25 crc kubenswrapper[4838]: I1207 09:48:25.581392 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"a37d598a4ad1f8e326405a22ca30de8c4dc862ce75ea17fb6791106fb71d0058"} Dec 07 09:48:29 crc kubenswrapper[4838]: I1207 09:48:29.645553 4838 generic.go:334] "Generic (PLEG): container finished" podID="9e6c35ea-c607-4c61-84d0-7bdf0d62add9" containerID="74d554e9726f99037638d695efda0e16eeb9beb22388e39db7871beca13265c6" exitCode=0 Dec 07 09:48:29 crc kubenswrapper[4838]: I1207 09:48:29.663584 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" event={"ID":"9e6c35ea-c607-4c61-84d0-7bdf0d62add9","Type":"ContainerDied","Data":"74d554e9726f99037638d695efda0e16eeb9beb22388e39db7871beca13265c6"} Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.112251 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.264335 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7676d\" (UniqueName: \"kubernetes.io/projected/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-kube-api-access-7676d\") pod \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.264430 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-inventory\") pod \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.264455 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ceph\") pod \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.264489 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ssh-key\") pod \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\" (UID: \"9e6c35ea-c607-4c61-84d0-7bdf0d62add9\") " Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.271651 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ceph" (OuterVolumeSpecName: "ceph") pod "9e6c35ea-c607-4c61-84d0-7bdf0d62add9" (UID: "9e6c35ea-c607-4c61-84d0-7bdf0d62add9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.281994 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-kube-api-access-7676d" (OuterVolumeSpecName: "kube-api-access-7676d") pod "9e6c35ea-c607-4c61-84d0-7bdf0d62add9" (UID: "9e6c35ea-c607-4c61-84d0-7bdf0d62add9"). InnerVolumeSpecName "kube-api-access-7676d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.294397 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-inventory" (OuterVolumeSpecName: "inventory") pod "9e6c35ea-c607-4c61-84d0-7bdf0d62add9" (UID: "9e6c35ea-c607-4c61-84d0-7bdf0d62add9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.302570 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9e6c35ea-c607-4c61-84d0-7bdf0d62add9" (UID: "9e6c35ea-c607-4c61-84d0-7bdf0d62add9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.366766 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.366810 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.366834 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.366843 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7676d\" (UniqueName: \"kubernetes.io/projected/9e6c35ea-c607-4c61-84d0-7bdf0d62add9-kube-api-access-7676d\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.680020 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" event={"ID":"9e6c35ea-c607-4c61-84d0-7bdf0d62add9","Type":"ContainerDied","Data":"4f87982f41ed74f6b44024f17e546e8f97e1d6bed1b20095b23082afa0030e9f"} Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.680062 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f87982f41ed74f6b44024f17e546e8f97e1d6bed1b20095b23082afa0030e9f" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.680121 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.769437 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq"] Dec 07 09:48:31 crc kubenswrapper[4838]: E1207 09:48:31.770285 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6c35ea-c607-4c61-84d0-7bdf0d62add9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.770309 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6c35ea-c607-4c61-84d0-7bdf0d62add9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.770528 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6c35ea-c607-4c61-84d0-7bdf0d62add9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.771258 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.775544 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.775628 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.775855 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.779028 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.779285 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.796020 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq"] Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.874255 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.874312 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.874394 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.874417 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmxjp\" (UniqueName: \"kubernetes.io/projected/687f9d1b-115c-4b7a-bd3c-35666e665ec3-kube-api-access-fmxjp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.976018 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.976064 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.976173 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.976207 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmxjp\" (UniqueName: \"kubernetes.io/projected/687f9d1b-115c-4b7a-bd3c-35666e665ec3-kube-api-access-fmxjp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.982529 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.982601 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.982833 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:31 crc kubenswrapper[4838]: I1207 09:48:31.992805 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmxjp\" (UniqueName: \"kubernetes.io/projected/687f9d1b-115c-4b7a-bd3c-35666e665ec3-kube-api-access-fmxjp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:32 crc kubenswrapper[4838]: I1207 09:48:32.094385 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:48:32 crc kubenswrapper[4838]: I1207 09:48:32.601504 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq"] Dec 07 09:48:32 crc kubenswrapper[4838]: W1207 09:48:32.610002 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod687f9d1b_115c_4b7a_bd3c_35666e665ec3.slice/crio-76bfd95ba46b9244e37aa0c76e84f2fac9b2ac4820853f4d886cbd1ca6c81003 WatchSource:0}: Error finding container 76bfd95ba46b9244e37aa0c76e84f2fac9b2ac4820853f4d886cbd1ca6c81003: Status 404 returned error can't find the container with id 76bfd95ba46b9244e37aa0c76e84f2fac9b2ac4820853f4d886cbd1ca6c81003 Dec 07 09:48:32 crc kubenswrapper[4838]: I1207 09:48:32.689544 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" event={"ID":"687f9d1b-115c-4b7a-bd3c-35666e665ec3","Type":"ContainerStarted","Data":"76bfd95ba46b9244e37aa0c76e84f2fac9b2ac4820853f4d886cbd1ca6c81003"} Dec 07 09:48:33 crc kubenswrapper[4838]: I1207 09:48:33.699955 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" event={"ID":"687f9d1b-115c-4b7a-bd3c-35666e665ec3","Type":"ContainerStarted","Data":"87a260d29d0dd2559f9cbd59a365765f5ec533774fa8de5c8fe11ba9a22ea2ae"} Dec 07 09:48:33 crc kubenswrapper[4838]: I1207 09:48:33.718904 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" podStartSLOduration=2.281002919 podStartE2EDuration="2.718885671s" podCreationTimestamp="2025-12-07 09:48:31 +0000 UTC" firstStartedPulling="2025-12-07 09:48:32.611342334 +0000 UTC m=+2529.318661351" lastFinishedPulling="2025-12-07 09:48:33.049225066 +0000 UTC m=+2529.756544103" observedRunningTime="2025-12-07 09:48:33.717834421 +0000 UTC m=+2530.425153448" watchObservedRunningTime="2025-12-07 09:48:33.718885671 +0000 UTC m=+2530.426204688" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.591231 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bc8gb"] Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.594006 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.601213 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bc8gb"] Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.627184 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-utilities\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.627481 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-catalog-content\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.627543 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f26hx\" (UniqueName: \"kubernetes.io/projected/7bef5aa4-3b9e-42c1-9c38-9ec073521083-kube-api-access-f26hx\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.729498 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-utilities\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.729546 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-catalog-content\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.729594 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f26hx\" (UniqueName: \"kubernetes.io/projected/7bef5aa4-3b9e-42c1-9c38-9ec073521083-kube-api-access-f26hx\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.730240 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-catalog-content\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.730242 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-utilities\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.764594 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f26hx\" (UniqueName: \"kubernetes.io/projected/7bef5aa4-3b9e-42c1-9c38-9ec073521083-kube-api-access-f26hx\") pod \"redhat-marketplace-bc8gb\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:38 crc kubenswrapper[4838]: I1207 09:48:38.927390 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:39 crc kubenswrapper[4838]: I1207 09:48:39.390192 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bc8gb"] Dec 07 09:48:39 crc kubenswrapper[4838]: W1207 09:48:39.412434 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bef5aa4_3b9e_42c1_9c38_9ec073521083.slice/crio-880e609ab7b855d3e94304ded0a1b7680476366dd8a2d6dc88da1c092cec0ba6 WatchSource:0}: Error finding container 880e609ab7b855d3e94304ded0a1b7680476366dd8a2d6dc88da1c092cec0ba6: Status 404 returned error can't find the container with id 880e609ab7b855d3e94304ded0a1b7680476366dd8a2d6dc88da1c092cec0ba6 Dec 07 09:48:39 crc kubenswrapper[4838]: I1207 09:48:39.751345 4838 generic.go:334] "Generic (PLEG): container finished" podID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerID="cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33" exitCode=0 Dec 07 09:48:39 crc kubenswrapper[4838]: I1207 09:48:39.751491 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bc8gb" event={"ID":"7bef5aa4-3b9e-42c1-9c38-9ec073521083","Type":"ContainerDied","Data":"cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33"} Dec 07 09:48:39 crc kubenswrapper[4838]: I1207 09:48:39.751529 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bc8gb" event={"ID":"7bef5aa4-3b9e-42c1-9c38-9ec073521083","Type":"ContainerStarted","Data":"880e609ab7b855d3e94304ded0a1b7680476366dd8a2d6dc88da1c092cec0ba6"} Dec 07 09:48:40 crc kubenswrapper[4838]: I1207 09:48:40.768505 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bc8gb" event={"ID":"7bef5aa4-3b9e-42c1-9c38-9ec073521083","Type":"ContainerStarted","Data":"1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c"} Dec 07 09:48:41 crc kubenswrapper[4838]: I1207 09:48:41.797327 4838 generic.go:334] "Generic (PLEG): container finished" podID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerID="1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c" exitCode=0 Dec 07 09:48:41 crc kubenswrapper[4838]: I1207 09:48:41.797421 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bc8gb" event={"ID":"7bef5aa4-3b9e-42c1-9c38-9ec073521083","Type":"ContainerDied","Data":"1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c"} Dec 07 09:48:42 crc kubenswrapper[4838]: I1207 09:48:42.808101 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bc8gb" event={"ID":"7bef5aa4-3b9e-42c1-9c38-9ec073521083","Type":"ContainerStarted","Data":"68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18"} Dec 07 09:48:42 crc kubenswrapper[4838]: I1207 09:48:42.840287 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bc8gb" podStartSLOduration=2.435973008 podStartE2EDuration="4.840262289s" podCreationTimestamp="2025-12-07 09:48:38 +0000 UTC" firstStartedPulling="2025-12-07 09:48:39.752829302 +0000 UTC m=+2536.460148319" lastFinishedPulling="2025-12-07 09:48:42.157118583 +0000 UTC m=+2538.864437600" observedRunningTime="2025-12-07 09:48:42.827185539 +0000 UTC m=+2539.534504586" watchObservedRunningTime="2025-12-07 09:48:42.840262289 +0000 UTC m=+2539.547581346" Dec 07 09:48:48 crc kubenswrapper[4838]: I1207 09:48:48.928637 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:48 crc kubenswrapper[4838]: I1207 09:48:48.929179 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:48 crc kubenswrapper[4838]: I1207 09:48:48.987584 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:49 crc kubenswrapper[4838]: I1207 09:48:49.927429 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:50 crc kubenswrapper[4838]: I1207 09:48:50.004996 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bc8gb"] Dec 07 09:48:51 crc kubenswrapper[4838]: I1207 09:48:51.886642 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bc8gb" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="registry-server" containerID="cri-o://68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18" gracePeriod=2 Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.371661 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.498006 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-catalog-content\") pod \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.498111 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f26hx\" (UniqueName: \"kubernetes.io/projected/7bef5aa4-3b9e-42c1-9c38-9ec073521083-kube-api-access-f26hx\") pod \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.498153 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-utilities\") pod \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\" (UID: \"7bef5aa4-3b9e-42c1-9c38-9ec073521083\") " Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.499888 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-utilities" (OuterVolumeSpecName: "utilities") pod "7bef5aa4-3b9e-42c1-9c38-9ec073521083" (UID: "7bef5aa4-3b9e-42c1-9c38-9ec073521083"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.514828 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bef5aa4-3b9e-42c1-9c38-9ec073521083-kube-api-access-f26hx" (OuterVolumeSpecName: "kube-api-access-f26hx") pod "7bef5aa4-3b9e-42c1-9c38-9ec073521083" (UID: "7bef5aa4-3b9e-42c1-9c38-9ec073521083"). InnerVolumeSpecName "kube-api-access-f26hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.529511 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bef5aa4-3b9e-42c1-9c38-9ec073521083" (UID: "7bef5aa4-3b9e-42c1-9c38-9ec073521083"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.601093 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.601144 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f26hx\" (UniqueName: \"kubernetes.io/projected/7bef5aa4-3b9e-42c1-9c38-9ec073521083-kube-api-access-f26hx\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.601167 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bef5aa4-3b9e-42c1-9c38-9ec073521083-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.895887 4838 generic.go:334] "Generic (PLEG): container finished" podID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerID="68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18" exitCode=0 Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.895932 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bc8gb" event={"ID":"7bef5aa4-3b9e-42c1-9c38-9ec073521083","Type":"ContainerDied","Data":"68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18"} Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.896435 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bc8gb" event={"ID":"7bef5aa4-3b9e-42c1-9c38-9ec073521083","Type":"ContainerDied","Data":"880e609ab7b855d3e94304ded0a1b7680476366dd8a2d6dc88da1c092cec0ba6"} Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.896457 4838 scope.go:117] "RemoveContainer" containerID="68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.896043 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bc8gb" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.923165 4838 scope.go:117] "RemoveContainer" containerID="1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c" Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.950047 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bc8gb"] Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.958773 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bc8gb"] Dec 07 09:48:52 crc kubenswrapper[4838]: I1207 09:48:52.970124 4838 scope.go:117] "RemoveContainer" containerID="cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33" Dec 07 09:48:53 crc kubenswrapper[4838]: I1207 09:48:53.026685 4838 scope.go:117] "RemoveContainer" containerID="68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18" Dec 07 09:48:53 crc kubenswrapper[4838]: E1207 09:48:53.027400 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18\": container with ID starting with 68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18 not found: ID does not exist" containerID="68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18" Dec 07 09:48:53 crc kubenswrapper[4838]: I1207 09:48:53.027430 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18"} err="failed to get container status \"68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18\": rpc error: code = NotFound desc = could not find container \"68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18\": container with ID starting with 68b2f7d1dfba4db8439c7153f4e510095ccab1378920fc5e7be691b75e483b18 not found: ID does not exist" Dec 07 09:48:53 crc kubenswrapper[4838]: I1207 09:48:53.027451 4838 scope.go:117] "RemoveContainer" containerID="1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c" Dec 07 09:48:53 crc kubenswrapper[4838]: E1207 09:48:53.027733 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c\": container with ID starting with 1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c not found: ID does not exist" containerID="1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c" Dec 07 09:48:53 crc kubenswrapper[4838]: I1207 09:48:53.027768 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c"} err="failed to get container status \"1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c\": rpc error: code = NotFound desc = could not find container \"1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c\": container with ID starting with 1b3d4e0092d042ea1750053fce2332371fe7d77cb4ec834a883a4e68bc49626c not found: ID does not exist" Dec 07 09:48:53 crc kubenswrapper[4838]: I1207 09:48:53.027791 4838 scope.go:117] "RemoveContainer" containerID="cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33" Dec 07 09:48:53 crc kubenswrapper[4838]: E1207 09:48:53.028324 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33\": container with ID starting with cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33 not found: ID does not exist" containerID="cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33" Dec 07 09:48:53 crc kubenswrapper[4838]: I1207 09:48:53.028348 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33"} err="failed to get container status \"cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33\": rpc error: code = NotFound desc = could not find container \"cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33\": container with ID starting with cafa0c71be2b13d3f3dadea870d390d346c7c7a3ab84f4c280bfb35d75d40d33 not found: ID does not exist" Dec 07 09:48:53 crc kubenswrapper[4838]: I1207 09:48:53.629249 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" path="/var/lib/kubelet/pods/7bef5aa4-3b9e-42c1-9c38-9ec073521083/volumes" Dec 07 09:49:27 crc kubenswrapper[4838]: I1207 09:49:27.194465 4838 generic.go:334] "Generic (PLEG): container finished" podID="687f9d1b-115c-4b7a-bd3c-35666e665ec3" containerID="87a260d29d0dd2559f9cbd59a365765f5ec533774fa8de5c8fe11ba9a22ea2ae" exitCode=0 Dec 07 09:49:27 crc kubenswrapper[4838]: I1207 09:49:27.194615 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" event={"ID":"687f9d1b-115c-4b7a-bd3c-35666e665ec3","Type":"ContainerDied","Data":"87a260d29d0dd2559f9cbd59a365765f5ec533774fa8de5c8fe11ba9a22ea2ae"} Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.582866 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.676254 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ceph\") pod \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.676580 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmxjp\" (UniqueName: \"kubernetes.io/projected/687f9d1b-115c-4b7a-bd3c-35666e665ec3-kube-api-access-fmxjp\") pod \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.676771 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ssh-key\") pod \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.676921 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-inventory\") pod \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\" (UID: \"687f9d1b-115c-4b7a-bd3c-35666e665ec3\") " Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.682147 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/687f9d1b-115c-4b7a-bd3c-35666e665ec3-kube-api-access-fmxjp" (OuterVolumeSpecName: "kube-api-access-fmxjp") pod "687f9d1b-115c-4b7a-bd3c-35666e665ec3" (UID: "687f9d1b-115c-4b7a-bd3c-35666e665ec3"). InnerVolumeSpecName "kube-api-access-fmxjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.682273 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ceph" (OuterVolumeSpecName: "ceph") pod "687f9d1b-115c-4b7a-bd3c-35666e665ec3" (UID: "687f9d1b-115c-4b7a-bd3c-35666e665ec3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.703976 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "687f9d1b-115c-4b7a-bd3c-35666e665ec3" (UID: "687f9d1b-115c-4b7a-bd3c-35666e665ec3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.713878 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-inventory" (OuterVolumeSpecName: "inventory") pod "687f9d1b-115c-4b7a-bd3c-35666e665ec3" (UID: "687f9d1b-115c-4b7a-bd3c-35666e665ec3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.778848 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.778874 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.778884 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/687f9d1b-115c-4b7a-bd3c-35666e665ec3-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:28 crc kubenswrapper[4838]: I1207 09:49:28.778893 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmxjp\" (UniqueName: \"kubernetes.io/projected/687f9d1b-115c-4b7a-bd3c-35666e665ec3-kube-api-access-fmxjp\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.215554 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" event={"ID":"687f9d1b-115c-4b7a-bd3c-35666e665ec3","Type":"ContainerDied","Data":"76bfd95ba46b9244e37aa0c76e84f2fac9b2ac4820853f4d886cbd1ca6c81003"} Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.215659 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76bfd95ba46b9244e37aa0c76e84f2fac9b2ac4820853f4d886cbd1ca6c81003" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.215600 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.322449 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8grpd"] Dec 07 09:49:29 crc kubenswrapper[4838]: E1207 09:49:29.323722 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="extract-utilities" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.323744 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="extract-utilities" Dec 07 09:49:29 crc kubenswrapper[4838]: E1207 09:49:29.323779 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="extract-content" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.323788 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="extract-content" Dec 07 09:49:29 crc kubenswrapper[4838]: E1207 09:49:29.323801 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="registry-server" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.323826 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="registry-server" Dec 07 09:49:29 crc kubenswrapper[4838]: E1207 09:49:29.323839 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687f9d1b-115c-4b7a-bd3c-35666e665ec3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.323849 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="687f9d1b-115c-4b7a-bd3c-35666e665ec3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.324237 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bef5aa4-3b9e-42c1-9c38-9ec073521083" containerName="registry-server" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.324279 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="687f9d1b-115c-4b7a-bd3c-35666e665ec3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.325021 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.329235 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.329559 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.329778 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.333272 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.333611 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.377264 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8grpd"] Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.494790 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.494892 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8k49\" (UniqueName: \"kubernetes.io/projected/131b49fd-19d7-4188-a131-2d684a0eee60-kube-api-access-m8k49\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.494928 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ceph\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.494962 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.596688 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.596755 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8k49\" (UniqueName: \"kubernetes.io/projected/131b49fd-19d7-4188-a131-2d684a0eee60-kube-api-access-m8k49\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.596796 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ceph\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.596895 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.602539 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.602694 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.604423 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ceph\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.615734 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8k49\" (UniqueName: \"kubernetes.io/projected/131b49fd-19d7-4188-a131-2d684a0eee60-kube-api-access-m8k49\") pod \"ssh-known-hosts-edpm-deployment-8grpd\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:29 crc kubenswrapper[4838]: I1207 09:49:29.650845 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:30 crc kubenswrapper[4838]: I1207 09:49:30.223231 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8grpd"] Dec 07 09:49:31 crc kubenswrapper[4838]: I1207 09:49:31.236040 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" event={"ID":"131b49fd-19d7-4188-a131-2d684a0eee60","Type":"ContainerStarted","Data":"2c7b1f871b6dd9c8390db60b3ff0876798802273d6bd879c9e12a1dd9d4d53b8"} Dec 07 09:49:31 crc kubenswrapper[4838]: I1207 09:49:31.236321 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" event={"ID":"131b49fd-19d7-4188-a131-2d684a0eee60","Type":"ContainerStarted","Data":"f84bebb139430eaf6f95c86248b917c93ab33867506f9642c4ec205419950ca6"} Dec 07 09:49:31 crc kubenswrapper[4838]: I1207 09:49:31.261016 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" podStartSLOduration=1.8374543349999999 podStartE2EDuration="2.26099494s" podCreationTimestamp="2025-12-07 09:49:29 +0000 UTC" firstStartedPulling="2025-12-07 09:49:30.228226406 +0000 UTC m=+2586.935545423" lastFinishedPulling="2025-12-07 09:49:30.651766981 +0000 UTC m=+2587.359086028" observedRunningTime="2025-12-07 09:49:31.253720219 +0000 UTC m=+2587.961039236" watchObservedRunningTime="2025-12-07 09:49:31.26099494 +0000 UTC m=+2587.968313957" Dec 07 09:49:41 crc kubenswrapper[4838]: I1207 09:49:41.359315 4838 generic.go:334] "Generic (PLEG): container finished" podID="131b49fd-19d7-4188-a131-2d684a0eee60" containerID="2c7b1f871b6dd9c8390db60b3ff0876798802273d6bd879c9e12a1dd9d4d53b8" exitCode=0 Dec 07 09:49:41 crc kubenswrapper[4838]: I1207 09:49:41.359387 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" event={"ID":"131b49fd-19d7-4188-a131-2d684a0eee60","Type":"ContainerDied","Data":"2c7b1f871b6dd9c8390db60b3ff0876798802273d6bd879c9e12a1dd9d4d53b8"} Dec 07 09:49:42 crc kubenswrapper[4838]: I1207 09:49:42.799295 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:42 crc kubenswrapper[4838]: I1207 09:49:42.979464 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8k49\" (UniqueName: \"kubernetes.io/projected/131b49fd-19d7-4188-a131-2d684a0eee60-kube-api-access-m8k49\") pod \"131b49fd-19d7-4188-a131-2d684a0eee60\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " Dec 07 09:49:42 crc kubenswrapper[4838]: I1207 09:49:42.979625 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ceph\") pod \"131b49fd-19d7-4188-a131-2d684a0eee60\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " Dec 07 09:49:42 crc kubenswrapper[4838]: I1207 09:49:42.979676 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ssh-key-openstack-edpm-ipam\") pod \"131b49fd-19d7-4188-a131-2d684a0eee60\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " Dec 07 09:49:42 crc kubenswrapper[4838]: I1207 09:49:42.979801 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-inventory-0\") pod \"131b49fd-19d7-4188-a131-2d684a0eee60\" (UID: \"131b49fd-19d7-4188-a131-2d684a0eee60\") " Dec 07 09:49:42 crc kubenswrapper[4838]: I1207 09:49:42.986890 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131b49fd-19d7-4188-a131-2d684a0eee60-kube-api-access-m8k49" (OuterVolumeSpecName: "kube-api-access-m8k49") pod "131b49fd-19d7-4188-a131-2d684a0eee60" (UID: "131b49fd-19d7-4188-a131-2d684a0eee60"). InnerVolumeSpecName "kube-api-access-m8k49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:49:42 crc kubenswrapper[4838]: I1207 09:49:42.994533 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ceph" (OuterVolumeSpecName: "ceph") pod "131b49fd-19d7-4188-a131-2d684a0eee60" (UID: "131b49fd-19d7-4188-a131-2d684a0eee60"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.014547 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "131b49fd-19d7-4188-a131-2d684a0eee60" (UID: "131b49fd-19d7-4188-a131-2d684a0eee60"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.027567 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "131b49fd-19d7-4188-a131-2d684a0eee60" (UID: "131b49fd-19d7-4188-a131-2d684a0eee60"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.082434 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8k49\" (UniqueName: \"kubernetes.io/projected/131b49fd-19d7-4188-a131-2d684a0eee60-kube-api-access-m8k49\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.082483 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.082497 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.082513 4838 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/131b49fd-19d7-4188-a131-2d684a0eee60-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.375838 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" event={"ID":"131b49fd-19d7-4188-a131-2d684a0eee60","Type":"ContainerDied","Data":"f84bebb139430eaf6f95c86248b917c93ab33867506f9642c4ec205419950ca6"} Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.375883 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f84bebb139430eaf6f95c86248b917c93ab33867506f9642c4ec205419950ca6" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.376386 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8grpd" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.477880 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5"] Dec 07 09:49:43 crc kubenswrapper[4838]: E1207 09:49:43.478414 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131b49fd-19d7-4188-a131-2d684a0eee60" containerName="ssh-known-hosts-edpm-deployment" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.478436 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="131b49fd-19d7-4188-a131-2d684a0eee60" containerName="ssh-known-hosts-edpm-deployment" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.478656 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="131b49fd-19d7-4188-a131-2d684a0eee60" containerName="ssh-known-hosts-edpm-deployment" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.479424 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.488528 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.489021 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.489163 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.497709 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.497961 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.539665 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5"] Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.610983 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.611147 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.611372 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwl7r\" (UniqueName: \"kubernetes.io/projected/e471497e-008f-40a2-929e-33e08ea45f63-kube-api-access-dwl7r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.611402 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.713588 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwl7r\" (UniqueName: \"kubernetes.io/projected/e471497e-008f-40a2-929e-33e08ea45f63-kube-api-access-dwl7r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.713656 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.713722 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.713803 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.717427 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.717779 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.717887 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.738618 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwl7r\" (UniqueName: \"kubernetes.io/projected/e471497e-008f-40a2-929e-33e08ea45f63-kube-api-access-dwl7r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d67k5\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:43 crc kubenswrapper[4838]: I1207 09:49:43.803456 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:44 crc kubenswrapper[4838]: I1207 09:49:44.362078 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5"] Dec 07 09:49:44 crc kubenswrapper[4838]: W1207 09:49:44.369213 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode471497e_008f_40a2_929e_33e08ea45f63.slice/crio-8dfb6e7448eb8410d6681260ce07a45983f4824696c9ab78fcf117def29905f7 WatchSource:0}: Error finding container 8dfb6e7448eb8410d6681260ce07a45983f4824696c9ab78fcf117def29905f7: Status 404 returned error can't find the container with id 8dfb6e7448eb8410d6681260ce07a45983f4824696c9ab78fcf117def29905f7 Dec 07 09:49:44 crc kubenswrapper[4838]: I1207 09:49:44.387688 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" event={"ID":"e471497e-008f-40a2-929e-33e08ea45f63","Type":"ContainerStarted","Data":"8dfb6e7448eb8410d6681260ce07a45983f4824696c9ab78fcf117def29905f7"} Dec 07 09:49:45 crc kubenswrapper[4838]: I1207 09:49:45.397305 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" event={"ID":"e471497e-008f-40a2-929e-33e08ea45f63","Type":"ContainerStarted","Data":"1e05119d0322d0cc8264dbd41b9b57e5f5fb5e0fdd8a088c1c8b47a6e6b6ea89"} Dec 07 09:49:45 crc kubenswrapper[4838]: I1207 09:49:45.422894 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" podStartSLOduration=2.024634921 podStartE2EDuration="2.422872232s" podCreationTimestamp="2025-12-07 09:49:43 +0000 UTC" firstStartedPulling="2025-12-07 09:49:44.374508438 +0000 UTC m=+2601.081827465" lastFinishedPulling="2025-12-07 09:49:44.772745759 +0000 UTC m=+2601.480064776" observedRunningTime="2025-12-07 09:49:45.420146944 +0000 UTC m=+2602.127466001" watchObservedRunningTime="2025-12-07 09:49:45.422872232 +0000 UTC m=+2602.130191259" Dec 07 09:49:53 crc kubenswrapper[4838]: I1207 09:49:53.467503 4838 generic.go:334] "Generic (PLEG): container finished" podID="e471497e-008f-40a2-929e-33e08ea45f63" containerID="1e05119d0322d0cc8264dbd41b9b57e5f5fb5e0fdd8a088c1c8b47a6e6b6ea89" exitCode=0 Dec 07 09:49:53 crc kubenswrapper[4838]: I1207 09:49:53.467607 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" event={"ID":"e471497e-008f-40a2-929e-33e08ea45f63","Type":"ContainerDied","Data":"1e05119d0322d0cc8264dbd41b9b57e5f5fb5e0fdd8a088c1c8b47a6e6b6ea89"} Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.895332 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.930539 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ceph\") pod \"e471497e-008f-40a2-929e-33e08ea45f63\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.930858 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-inventory\") pod \"e471497e-008f-40a2-929e-33e08ea45f63\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.930996 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwl7r\" (UniqueName: \"kubernetes.io/projected/e471497e-008f-40a2-929e-33e08ea45f63-kube-api-access-dwl7r\") pod \"e471497e-008f-40a2-929e-33e08ea45f63\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.931087 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ssh-key\") pod \"e471497e-008f-40a2-929e-33e08ea45f63\" (UID: \"e471497e-008f-40a2-929e-33e08ea45f63\") " Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.936264 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ceph" (OuterVolumeSpecName: "ceph") pod "e471497e-008f-40a2-929e-33e08ea45f63" (UID: "e471497e-008f-40a2-929e-33e08ea45f63"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.939966 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e471497e-008f-40a2-929e-33e08ea45f63-kube-api-access-dwl7r" (OuterVolumeSpecName: "kube-api-access-dwl7r") pod "e471497e-008f-40a2-929e-33e08ea45f63" (UID: "e471497e-008f-40a2-929e-33e08ea45f63"). InnerVolumeSpecName "kube-api-access-dwl7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.960874 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e471497e-008f-40a2-929e-33e08ea45f63" (UID: "e471497e-008f-40a2-929e-33e08ea45f63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:54 crc kubenswrapper[4838]: I1207 09:49:54.972035 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-inventory" (OuterVolumeSpecName: "inventory") pod "e471497e-008f-40a2-929e-33e08ea45f63" (UID: "e471497e-008f-40a2-929e-33e08ea45f63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.032100 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.032134 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwl7r\" (UniqueName: \"kubernetes.io/projected/e471497e-008f-40a2-929e-33e08ea45f63-kube-api-access-dwl7r\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.032144 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.032152 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e471497e-008f-40a2-929e-33e08ea45f63-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.488487 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" event={"ID":"e471497e-008f-40a2-929e-33e08ea45f63","Type":"ContainerDied","Data":"8dfb6e7448eb8410d6681260ce07a45983f4824696c9ab78fcf117def29905f7"} Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.488548 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dfb6e7448eb8410d6681260ce07a45983f4824696c9ab78fcf117def29905f7" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.488591 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d67k5" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.635609 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g"] Dec 07 09:49:55 crc kubenswrapper[4838]: E1207 09:49:55.636746 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e471497e-008f-40a2-929e-33e08ea45f63" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.636789 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e471497e-008f-40a2-929e-33e08ea45f63" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.637237 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e471497e-008f-40a2-929e-33e08ea45f63" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.638494 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.643586 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g"] Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.644103 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.644272 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.644647 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.645794 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.646281 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.746758 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.746836 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv2w9\" (UniqueName: \"kubernetes.io/projected/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-kube-api-access-hv2w9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.746894 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.746951 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.848063 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.848110 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv2w9\" (UniqueName: \"kubernetes.io/projected/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-kube-api-access-hv2w9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.848148 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.848185 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.852622 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.852705 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.852962 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.873799 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv2w9\" (UniqueName: \"kubernetes.io/projected/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-kube-api-access-hv2w9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:55 crc kubenswrapper[4838]: I1207 09:49:55.977188 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:49:56 crc kubenswrapper[4838]: I1207 09:49:56.509875 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g"] Dec 07 09:49:57 crc kubenswrapper[4838]: I1207 09:49:57.536980 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" event={"ID":"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa","Type":"ContainerStarted","Data":"2a7f45c686edcaad10dc8a356a726ed9f84dbd1ea45159eed163d12a1641efe2"} Dec 07 09:49:57 crc kubenswrapper[4838]: I1207 09:49:57.537357 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" event={"ID":"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa","Type":"ContainerStarted","Data":"fc89f39408ab9626ead0070c7c53357924d7547d2a6d848e49e1e920f05fdfb7"} Dec 07 09:49:57 crc kubenswrapper[4838]: I1207 09:49:57.573182 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" podStartSLOduration=2.174382549 podStartE2EDuration="2.573161425s" podCreationTimestamp="2025-12-07 09:49:55 +0000 UTC" firstStartedPulling="2025-12-07 09:49:56.526887522 +0000 UTC m=+2613.234206559" lastFinishedPulling="2025-12-07 09:49:56.925666418 +0000 UTC m=+2613.632985435" observedRunningTime="2025-12-07 09:49:57.562712166 +0000 UTC m=+2614.270031183" watchObservedRunningTime="2025-12-07 09:49:57.573161425 +0000 UTC m=+2614.280480452" Dec 07 09:50:07 crc kubenswrapper[4838]: I1207 09:50:07.620078 4838 generic.go:334] "Generic (PLEG): container finished" podID="df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" containerID="2a7f45c686edcaad10dc8a356a726ed9f84dbd1ea45159eed163d12a1641efe2" exitCode=0 Dec 07 09:50:07 crc kubenswrapper[4838]: I1207 09:50:07.631838 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" event={"ID":"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa","Type":"ContainerDied","Data":"2a7f45c686edcaad10dc8a356a726ed9f84dbd1ea45159eed163d12a1641efe2"} Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.039579 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.102181 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ceph\") pod \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.102252 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv2w9\" (UniqueName: \"kubernetes.io/projected/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-kube-api-access-hv2w9\") pod \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.102294 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-inventory\") pod \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.102321 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ssh-key\") pod \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\" (UID: \"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa\") " Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.111845 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ceph" (OuterVolumeSpecName: "ceph") pod "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" (UID: "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.135869 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-kube-api-access-hv2w9" (OuterVolumeSpecName: "kube-api-access-hv2w9") pod "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" (UID: "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa"). InnerVolumeSpecName "kube-api-access-hv2w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.141008 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" (UID: "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.157450 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-inventory" (OuterVolumeSpecName: "inventory") pod "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" (UID: "df0eb7df-51e9-4da6-9e69-0b11a79ad3fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.204598 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.204651 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.204665 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv2w9\" (UniqueName: \"kubernetes.io/projected/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-kube-api-access-hv2w9\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.204683 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df0eb7df-51e9-4da6-9e69-0b11a79ad3fa-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.651007 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" event={"ID":"df0eb7df-51e9-4da6-9e69-0b11a79ad3fa","Type":"ContainerDied","Data":"fc89f39408ab9626ead0070c7c53357924d7547d2a6d848e49e1e920f05fdfb7"} Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.651564 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc89f39408ab9626ead0070c7c53357924d7547d2a6d848e49e1e920f05fdfb7" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.651116 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.756111 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s"] Dec 07 09:50:09 crc kubenswrapper[4838]: E1207 09:50:09.756464 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.756486 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.756778 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="df0eb7df-51e9-4da6-9e69-0b11a79ad3fa" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.757421 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.762226 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.763153 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.763617 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.765916 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.766138 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.766171 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.766653 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.766952 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.784174 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s"] Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.814221 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfk9j\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-kube-api-access-rfk9j\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.814493 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.814703 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.814867 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.815039 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.815259 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.815428 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.815618 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.815765 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.815963 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.816121 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.816289 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.816463 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918058 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfk9j\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-kube-api-access-rfk9j\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918103 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918145 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918169 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918200 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918235 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918255 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918282 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918300 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918321 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918341 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918358 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.918392 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.924701 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.928769 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.929296 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.929754 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.930081 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.930280 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.930876 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.931930 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.932834 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.933645 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.933744 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.935599 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:09 crc kubenswrapper[4838]: I1207 09:50:09.937376 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfk9j\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-kube-api-access-rfk9j\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:10 crc kubenswrapper[4838]: I1207 09:50:10.075209 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:10 crc kubenswrapper[4838]: I1207 09:50:10.669373 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s"] Dec 07 09:50:11 crc kubenswrapper[4838]: I1207 09:50:11.672068 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" event={"ID":"84df8337-c690-4c33-9d82-da53b4de5be5","Type":"ContainerStarted","Data":"554eb1789d5802fef7d51c7dc0fc8cb6f48a0674e5fb820c7b870233bf3d8b82"} Dec 07 09:50:11 crc kubenswrapper[4838]: I1207 09:50:11.672723 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" event={"ID":"84df8337-c690-4c33-9d82-da53b4de5be5","Type":"ContainerStarted","Data":"7ebde76353d706d84f35296fc128ce491079ecfbb5f4293e9f0121638d9c7e8c"} Dec 07 09:50:11 crc kubenswrapper[4838]: I1207 09:50:11.705236 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" podStartSLOduration=2.287389043 podStartE2EDuration="2.705219065s" podCreationTimestamp="2025-12-07 09:50:09 +0000 UTC" firstStartedPulling="2025-12-07 09:50:10.686550632 +0000 UTC m=+2627.393869659" lastFinishedPulling="2025-12-07 09:50:11.104380624 +0000 UTC m=+2627.811699681" observedRunningTime="2025-12-07 09:50:11.702266111 +0000 UTC m=+2628.409585218" watchObservedRunningTime="2025-12-07 09:50:11.705219065 +0000 UTC m=+2628.412538082" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.390660 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9csbz"] Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.392880 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.402037 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9csbz"] Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.411343 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2p59\" (UniqueName: \"kubernetes.io/projected/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-kube-api-access-c2p59\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.411387 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-catalog-content\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.411610 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-utilities\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.513019 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-utilities\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.513082 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2p59\" (UniqueName: \"kubernetes.io/projected/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-kube-api-access-c2p59\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.513107 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-catalog-content\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.513604 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-utilities\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.513615 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-catalog-content\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.531233 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2p59\" (UniqueName: \"kubernetes.io/projected/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-kube-api-access-c2p59\") pod \"redhat-operators-9csbz\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:14 crc kubenswrapper[4838]: I1207 09:50:14.729854 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:15 crc kubenswrapper[4838]: I1207 09:50:15.247825 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9csbz"] Dec 07 09:50:15 crc kubenswrapper[4838]: I1207 09:50:15.701709 4838 generic.go:334] "Generic (PLEG): container finished" podID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerID="46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56" exitCode=0 Dec 07 09:50:15 crc kubenswrapper[4838]: I1207 09:50:15.702883 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9csbz" event={"ID":"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d","Type":"ContainerDied","Data":"46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56"} Dec 07 09:50:15 crc kubenswrapper[4838]: I1207 09:50:15.702989 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9csbz" event={"ID":"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d","Type":"ContainerStarted","Data":"d482ea2c303075aed30c50f0cf732837afa1e8e62d4ecd4327287123824a0755"} Dec 07 09:50:16 crc kubenswrapper[4838]: I1207 09:50:16.714355 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9csbz" event={"ID":"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d","Type":"ContainerStarted","Data":"62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba"} Dec 07 09:50:19 crc kubenswrapper[4838]: I1207 09:50:19.741115 4838 generic.go:334] "Generic (PLEG): container finished" podID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerID="62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba" exitCode=0 Dec 07 09:50:19 crc kubenswrapper[4838]: I1207 09:50:19.741192 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9csbz" event={"ID":"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d","Type":"ContainerDied","Data":"62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba"} Dec 07 09:50:20 crc kubenswrapper[4838]: I1207 09:50:20.754752 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9csbz" event={"ID":"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d","Type":"ContainerStarted","Data":"f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07"} Dec 07 09:50:20 crc kubenswrapper[4838]: I1207 09:50:20.774713 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9csbz" podStartSLOduration=2.35494211 podStartE2EDuration="6.774696969s" podCreationTimestamp="2025-12-07 09:50:14 +0000 UTC" firstStartedPulling="2025-12-07 09:50:15.704517439 +0000 UTC m=+2632.411836456" lastFinishedPulling="2025-12-07 09:50:20.124272288 +0000 UTC m=+2636.831591315" observedRunningTime="2025-12-07 09:50:20.773119034 +0000 UTC m=+2637.480438051" watchObservedRunningTime="2025-12-07 09:50:20.774696969 +0000 UTC m=+2637.482015986" Dec 07 09:50:24 crc kubenswrapper[4838]: I1207 09:50:24.492897 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:50:24 crc kubenswrapper[4838]: I1207 09:50:24.493664 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:50:24 crc kubenswrapper[4838]: I1207 09:50:24.731271 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:24 crc kubenswrapper[4838]: I1207 09:50:24.732362 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:25 crc kubenswrapper[4838]: I1207 09:50:25.783796 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9csbz" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="registry-server" probeResult="failure" output=< Dec 07 09:50:25 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 09:50:25 crc kubenswrapper[4838]: > Dec 07 09:50:34 crc kubenswrapper[4838]: I1207 09:50:34.779380 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:34 crc kubenswrapper[4838]: I1207 09:50:34.831935 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:35 crc kubenswrapper[4838]: I1207 09:50:35.013743 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9csbz"] Dec 07 09:50:35 crc kubenswrapper[4838]: I1207 09:50:35.879484 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9csbz" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="registry-server" containerID="cri-o://f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07" gracePeriod=2 Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.311728 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.412015 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-utilities\") pod \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.412549 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2p59\" (UniqueName: \"kubernetes.io/projected/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-kube-api-access-c2p59\") pod \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.412617 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-catalog-content\") pod \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\" (UID: \"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d\") " Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.413578 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-utilities" (OuterVolumeSpecName: "utilities") pod "a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" (UID: "a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.421655 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-kube-api-access-c2p59" (OuterVolumeSpecName: "kube-api-access-c2p59") pod "a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" (UID: "a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d"). InnerVolumeSpecName "kube-api-access-c2p59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.514696 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2p59\" (UniqueName: \"kubernetes.io/projected/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-kube-api-access-c2p59\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.514721 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.524130 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" (UID: "a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.616479 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.894017 4838 generic.go:334] "Generic (PLEG): container finished" podID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerID="f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07" exitCode=0 Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.894058 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9csbz" event={"ID":"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d","Type":"ContainerDied","Data":"f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07"} Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.894085 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9csbz" event={"ID":"a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d","Type":"ContainerDied","Data":"d482ea2c303075aed30c50f0cf732837afa1e8e62d4ecd4327287123824a0755"} Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.894099 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9csbz" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.894107 4838 scope.go:117] "RemoveContainer" containerID="f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.930417 4838 scope.go:117] "RemoveContainer" containerID="62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba" Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.933257 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9csbz"] Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.942214 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9csbz"] Dec 07 09:50:36 crc kubenswrapper[4838]: I1207 09:50:36.960006 4838 scope.go:117] "RemoveContainer" containerID="46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56" Dec 07 09:50:37 crc kubenswrapper[4838]: I1207 09:50:37.002164 4838 scope.go:117] "RemoveContainer" containerID="f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07" Dec 07 09:50:37 crc kubenswrapper[4838]: E1207 09:50:37.002733 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07\": container with ID starting with f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07 not found: ID does not exist" containerID="f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07" Dec 07 09:50:37 crc kubenswrapper[4838]: I1207 09:50:37.002772 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07"} err="failed to get container status \"f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07\": rpc error: code = NotFound desc = could not find container \"f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07\": container with ID starting with f77e63ddeded6a55d4a6fefedab978c1f468de71c937756edad05b7c61c7de07 not found: ID does not exist" Dec 07 09:50:37 crc kubenswrapper[4838]: I1207 09:50:37.002796 4838 scope.go:117] "RemoveContainer" containerID="62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba" Dec 07 09:50:37 crc kubenswrapper[4838]: E1207 09:50:37.003136 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba\": container with ID starting with 62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba not found: ID does not exist" containerID="62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba" Dec 07 09:50:37 crc kubenswrapper[4838]: I1207 09:50:37.003181 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba"} err="failed to get container status \"62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba\": rpc error: code = NotFound desc = could not find container \"62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba\": container with ID starting with 62cbf85df8d98d09007bd9f2c771bb653a93df502cafb2c4c01dc932c8dd9fba not found: ID does not exist" Dec 07 09:50:37 crc kubenswrapper[4838]: I1207 09:50:37.003211 4838 scope.go:117] "RemoveContainer" containerID="46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56" Dec 07 09:50:37 crc kubenswrapper[4838]: E1207 09:50:37.003931 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56\": container with ID starting with 46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56 not found: ID does not exist" containerID="46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56" Dec 07 09:50:37 crc kubenswrapper[4838]: I1207 09:50:37.003965 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56"} err="failed to get container status \"46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56\": rpc error: code = NotFound desc = could not find container \"46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56\": container with ID starting with 46249f0f54aa7e864fb563d624cd3227b1c61d9c07d2ec033981be1837949d56 not found: ID does not exist" Dec 07 09:50:37 crc kubenswrapper[4838]: I1207 09:50:37.627058 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" path="/var/lib/kubelet/pods/a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d/volumes" Dec 07 09:50:46 crc kubenswrapper[4838]: I1207 09:50:46.997153 4838 generic.go:334] "Generic (PLEG): container finished" podID="84df8337-c690-4c33-9d82-da53b4de5be5" containerID="554eb1789d5802fef7d51c7dc0fc8cb6f48a0674e5fb820c7b870233bf3d8b82" exitCode=0 Dec 07 09:50:46 crc kubenswrapper[4838]: I1207 09:50:46.997372 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" event={"ID":"84df8337-c690-4c33-9d82-da53b4de5be5","Type":"ContainerDied","Data":"554eb1789d5802fef7d51c7dc0fc8cb6f48a0674e5fb820c7b870233bf3d8b82"} Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.471698 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580360 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-bootstrap-combined-ca-bundle\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580404 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-neutron-metadata-combined-ca-bundle\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580426 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580453 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580670 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-libvirt-combined-ca-bundle\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580694 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfk9j\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-kube-api-access-rfk9j\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580713 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580750 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ceph\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580789 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ovn-combined-ca-bundle\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580845 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-nova-combined-ca-bundle\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580866 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-repo-setup-combined-ca-bundle\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580885 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-inventory\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.580936 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ssh-key\") pod \"84df8337-c690-4c33-9d82-da53b4de5be5\" (UID: \"84df8337-c690-4c33-9d82-da53b4de5be5\") " Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.589716 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.590379 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.590390 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.590940 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.590477 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-kube-api-access-rfk9j" (OuterVolumeSpecName: "kube-api-access-rfk9j") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "kube-api-access-rfk9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.590748 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.591275 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.591426 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.592392 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ceph" (OuterVolumeSpecName: "ceph") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.601041 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.607509 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.611807 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.614251 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-inventory" (OuterVolumeSpecName: "inventory") pod "84df8337-c690-4c33-9d82-da53b4de5be5" (UID: "84df8337-c690-4c33-9d82-da53b4de5be5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682459 4838 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682490 4838 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682499 4838 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682509 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682518 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682527 4838 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682537 4838 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682547 4838 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682556 4838 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682565 4838 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682575 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfk9j\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-kube-api-access-rfk9j\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682584 4838 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/84df8337-c690-4c33-9d82-da53b4de5be5-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:48 crc kubenswrapper[4838]: I1207 09:50:48.682593 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/84df8337-c690-4c33-9d82-da53b4de5be5-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.018039 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" event={"ID":"84df8337-c690-4c33-9d82-da53b4de5be5","Type":"ContainerDied","Data":"7ebde76353d706d84f35296fc128ce491079ecfbb5f4293e9f0121638d9c7e8c"} Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.018548 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ebde76353d706d84f35296fc128ce491079ecfbb5f4293e9f0121638d9c7e8c" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.018302 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.125961 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d"] Dec 07 09:50:49 crc kubenswrapper[4838]: E1207 09:50:49.126310 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84df8337-c690-4c33-9d82-da53b4de5be5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.126328 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="84df8337-c690-4c33-9d82-da53b4de5be5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:49 crc kubenswrapper[4838]: E1207 09:50:49.126341 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="registry-server" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.126349 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="registry-server" Dec 07 09:50:49 crc kubenswrapper[4838]: E1207 09:50:49.126367 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="extract-utilities" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.126375 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="extract-utilities" Dec 07 09:50:49 crc kubenswrapper[4838]: E1207 09:50:49.126394 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="extract-content" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.126400 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="extract-content" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.126569 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8f2fa83-6ceb-41c1-b7d6-dee6adcb946d" containerName="registry-server" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.126591 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="84df8337-c690-4c33-9d82-da53b4de5be5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.127158 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.129680 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.129700 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.130047 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.130109 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.131383 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.143850 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d"] Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.293698 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.293900 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.293925 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.293962 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9jtd\" (UniqueName: \"kubernetes.io/projected/d681f081-977d-48c0-ac82-ff3c61647de8-kube-api-access-c9jtd\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.395808 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.395888 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.395945 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9jtd\" (UniqueName: \"kubernetes.io/projected/d681f081-977d-48c0-ac82-ff3c61647de8-kube-api-access-c9jtd\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.396357 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.400541 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.408433 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.422378 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9jtd\" (UniqueName: \"kubernetes.io/projected/d681f081-977d-48c0-ac82-ff3c61647de8-kube-api-access-c9jtd\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.427370 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.445006 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:49 crc kubenswrapper[4838]: I1207 09:50:49.992924 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d"] Dec 07 09:50:50 crc kubenswrapper[4838]: I1207 09:50:50.028155 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" event={"ID":"d681f081-977d-48c0-ac82-ff3c61647de8","Type":"ContainerStarted","Data":"7a78f5031d6f9271940189bfffc82ecbac85f2fc783e3e00a28a667c11108ffb"} Dec 07 09:50:51 crc kubenswrapper[4838]: I1207 09:50:51.037683 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" event={"ID":"d681f081-977d-48c0-ac82-ff3c61647de8","Type":"ContainerStarted","Data":"36adc267abf342ed1b50d78fb7ad0da344b2a3c93e480d2da49c40cd1a747f8a"} Dec 07 09:50:51 crc kubenswrapper[4838]: I1207 09:50:51.063616 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" podStartSLOduration=1.592097385 podStartE2EDuration="2.063594823s" podCreationTimestamp="2025-12-07 09:50:49 +0000 UTC" firstStartedPulling="2025-12-07 09:50:50.010573427 +0000 UTC m=+2666.717892444" lastFinishedPulling="2025-12-07 09:50:50.482070865 +0000 UTC m=+2667.189389882" observedRunningTime="2025-12-07 09:50:51.062910784 +0000 UTC m=+2667.770229801" watchObservedRunningTime="2025-12-07 09:50:51.063594823 +0000 UTC m=+2667.770913840" Dec 07 09:50:54 crc kubenswrapper[4838]: I1207 09:50:54.493343 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:50:54 crc kubenswrapper[4838]: I1207 09:50:54.493880 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:50:56 crc kubenswrapper[4838]: E1207 09:50:56.415900 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd681f081_977d_48c0_ac82_ff3c61647de8.slice/crio-36adc267abf342ed1b50d78fb7ad0da344b2a3c93e480d2da49c40cd1a747f8a.scope\": RecentStats: unable to find data in memory cache]" Dec 07 09:50:57 crc kubenswrapper[4838]: I1207 09:50:57.086738 4838 generic.go:334] "Generic (PLEG): container finished" podID="d681f081-977d-48c0-ac82-ff3c61647de8" containerID="36adc267abf342ed1b50d78fb7ad0da344b2a3c93e480d2da49c40cd1a747f8a" exitCode=0 Dec 07 09:50:57 crc kubenswrapper[4838]: I1207 09:50:57.086856 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" event={"ID":"d681f081-977d-48c0-ac82-ff3c61647de8","Type":"ContainerDied","Data":"36adc267abf342ed1b50d78fb7ad0da344b2a3c93e480d2da49c40cd1a747f8a"} Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.496013 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.575200 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-inventory\") pod \"d681f081-977d-48c0-ac82-ff3c61647de8\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.575243 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ceph\") pod \"d681f081-977d-48c0-ac82-ff3c61647de8\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.575277 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ssh-key\") pod \"d681f081-977d-48c0-ac82-ff3c61647de8\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.575478 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9jtd\" (UniqueName: \"kubernetes.io/projected/d681f081-977d-48c0-ac82-ff3c61647de8-kube-api-access-c9jtd\") pod \"d681f081-977d-48c0-ac82-ff3c61647de8\" (UID: \"d681f081-977d-48c0-ac82-ff3c61647de8\") " Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.581142 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ceph" (OuterVolumeSpecName: "ceph") pod "d681f081-977d-48c0-ac82-ff3c61647de8" (UID: "d681f081-977d-48c0-ac82-ff3c61647de8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.582786 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d681f081-977d-48c0-ac82-ff3c61647de8-kube-api-access-c9jtd" (OuterVolumeSpecName: "kube-api-access-c9jtd") pod "d681f081-977d-48c0-ac82-ff3c61647de8" (UID: "d681f081-977d-48c0-ac82-ff3c61647de8"). InnerVolumeSpecName "kube-api-access-c9jtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.604074 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d681f081-977d-48c0-ac82-ff3c61647de8" (UID: "d681f081-977d-48c0-ac82-ff3c61647de8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.615966 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-inventory" (OuterVolumeSpecName: "inventory") pod "d681f081-977d-48c0-ac82-ff3c61647de8" (UID: "d681f081-977d-48c0-ac82-ff3c61647de8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.677789 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.678139 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.678214 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d681f081-977d-48c0-ac82-ff3c61647de8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:58 crc kubenswrapper[4838]: I1207 09:50:58.678285 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9jtd\" (UniqueName: \"kubernetes.io/projected/d681f081-977d-48c0-ac82-ff3c61647de8-kube-api-access-c9jtd\") on node \"crc\" DevicePath \"\"" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.108492 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" event={"ID":"d681f081-977d-48c0-ac82-ff3c61647de8","Type":"ContainerDied","Data":"7a78f5031d6f9271940189bfffc82ecbac85f2fc783e3e00a28a667c11108ffb"} Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.108541 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a78f5031d6f9271940189bfffc82ecbac85f2fc783e3e00a28a667c11108ffb" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.108608 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.250658 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c"] Dec 07 09:50:59 crc kubenswrapper[4838]: E1207 09:50:59.250992 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d681f081-977d-48c0-ac82-ff3c61647de8" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.251010 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d681f081-977d-48c0-ac82-ff3c61647de8" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.251172 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d681f081-977d-48c0-ac82-ff3c61647de8" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.251712 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.257500 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.257552 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.257916 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.260040 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.260248 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.262245 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.280125 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c"] Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.390934 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.391006 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.391042 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.391056 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.391096 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.391131 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfznh\" (UniqueName: \"kubernetes.io/projected/6ae7a9e2-889e-499b-8f64-d38976e8d429-kube-api-access-jfznh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.492431 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.492478 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.492533 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.492575 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfznh\" (UniqueName: \"kubernetes.io/projected/6ae7a9e2-889e-499b-8f64-d38976e8d429-kube-api-access-jfznh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.492641 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.492682 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.494700 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.496400 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.497653 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.499260 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.499959 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.512055 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfznh\" (UniqueName: \"kubernetes.io/projected/6ae7a9e2-889e-499b-8f64-d38976e8d429-kube-api-access-jfznh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-84r6c\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:50:59 crc kubenswrapper[4838]: I1207 09:50:59.569297 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:51:00 crc kubenswrapper[4838]: I1207 09:51:00.230271 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c"] Dec 07 09:51:01 crc kubenswrapper[4838]: I1207 09:51:01.126938 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" event={"ID":"6ae7a9e2-889e-499b-8f64-d38976e8d429","Type":"ContainerStarted","Data":"6ade5b3437125619a0c0b910e49873147b33db3993bfc2979215a7e054b22ff0"} Dec 07 09:51:01 crc kubenswrapper[4838]: I1207 09:51:01.127553 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" event={"ID":"6ae7a9e2-889e-499b-8f64-d38976e8d429","Type":"ContainerStarted","Data":"222cc699676fee880c44797b8f3cb1684a4cdce0789594cb435c7f0bc3c0b104"} Dec 07 09:51:01 crc kubenswrapper[4838]: I1207 09:51:01.155003 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" podStartSLOduration=1.747586879 podStartE2EDuration="2.154981823s" podCreationTimestamp="2025-12-07 09:50:59 +0000 UTC" firstStartedPulling="2025-12-07 09:51:00.226754079 +0000 UTC m=+2676.934073096" lastFinishedPulling="2025-12-07 09:51:00.634149013 +0000 UTC m=+2677.341468040" observedRunningTime="2025-12-07 09:51:01.14824371 +0000 UTC m=+2677.855562727" watchObservedRunningTime="2025-12-07 09:51:01.154981823 +0000 UTC m=+2677.862300840" Dec 07 09:51:24 crc kubenswrapper[4838]: I1207 09:51:24.493284 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:51:24 crc kubenswrapper[4838]: I1207 09:51:24.493880 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:51:24 crc kubenswrapper[4838]: I1207 09:51:24.493924 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:51:24 crc kubenswrapper[4838]: I1207 09:51:24.494676 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a37d598a4ad1f8e326405a22ca30de8c4dc862ce75ea17fb6791106fb71d0058"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:51:24 crc kubenswrapper[4838]: I1207 09:51:24.494735 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://a37d598a4ad1f8e326405a22ca30de8c4dc862ce75ea17fb6791106fb71d0058" gracePeriod=600 Dec 07 09:51:25 crc kubenswrapper[4838]: I1207 09:51:25.353978 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="a37d598a4ad1f8e326405a22ca30de8c4dc862ce75ea17fb6791106fb71d0058" exitCode=0 Dec 07 09:51:25 crc kubenswrapper[4838]: I1207 09:51:25.354077 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"a37d598a4ad1f8e326405a22ca30de8c4dc862ce75ea17fb6791106fb71d0058"} Dec 07 09:51:25 crc kubenswrapper[4838]: I1207 09:51:25.354451 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f"} Dec 07 09:51:25 crc kubenswrapper[4838]: I1207 09:51:25.354471 4838 scope.go:117] "RemoveContainer" containerID="1da6e35ca5eae87ad4b3e91b92a552c5c4c7365dedc1d557cde62b5fc4b972c8" Dec 07 09:52:27 crc kubenswrapper[4838]: I1207 09:52:27.930248 4838 generic.go:334] "Generic (PLEG): container finished" podID="6ae7a9e2-889e-499b-8f64-d38976e8d429" containerID="6ade5b3437125619a0c0b910e49873147b33db3993bfc2979215a7e054b22ff0" exitCode=0 Dec 07 09:52:27 crc kubenswrapper[4838]: I1207 09:52:27.930336 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" event={"ID":"6ae7a9e2-889e-499b-8f64-d38976e8d429","Type":"ContainerDied","Data":"6ade5b3437125619a0c0b910e49873147b33db3993bfc2979215a7e054b22ff0"} Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.347251 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.516310 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ssh-key\") pod \"6ae7a9e2-889e-499b-8f64-d38976e8d429\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.516713 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ceph\") pod \"6ae7a9e2-889e-499b-8f64-d38976e8d429\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.516752 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovncontroller-config-0\") pod \"6ae7a9e2-889e-499b-8f64-d38976e8d429\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.516786 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfznh\" (UniqueName: \"kubernetes.io/projected/6ae7a9e2-889e-499b-8f64-d38976e8d429-kube-api-access-jfznh\") pod \"6ae7a9e2-889e-499b-8f64-d38976e8d429\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.516867 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-inventory\") pod \"6ae7a9e2-889e-499b-8f64-d38976e8d429\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.516907 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovn-combined-ca-bundle\") pod \"6ae7a9e2-889e-499b-8f64-d38976e8d429\" (UID: \"6ae7a9e2-889e-499b-8f64-d38976e8d429\") " Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.521470 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6ae7a9e2-889e-499b-8f64-d38976e8d429" (UID: "6ae7a9e2-889e-499b-8f64-d38976e8d429"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.521534 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ceph" (OuterVolumeSpecName: "ceph") pod "6ae7a9e2-889e-499b-8f64-d38976e8d429" (UID: "6ae7a9e2-889e-499b-8f64-d38976e8d429"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.524128 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae7a9e2-889e-499b-8f64-d38976e8d429-kube-api-access-jfznh" (OuterVolumeSpecName: "kube-api-access-jfznh") pod "6ae7a9e2-889e-499b-8f64-d38976e8d429" (UID: "6ae7a9e2-889e-499b-8f64-d38976e8d429"). InnerVolumeSpecName "kube-api-access-jfznh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.543515 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6ae7a9e2-889e-499b-8f64-d38976e8d429" (UID: "6ae7a9e2-889e-499b-8f64-d38976e8d429"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.550171 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-inventory" (OuterVolumeSpecName: "inventory") pod "6ae7a9e2-889e-499b-8f64-d38976e8d429" (UID: "6ae7a9e2-889e-499b-8f64-d38976e8d429"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.555129 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6ae7a9e2-889e-499b-8f64-d38976e8d429" (UID: "6ae7a9e2-889e-499b-8f64-d38976e8d429"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.619230 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.619256 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.619265 4838 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.619279 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfznh\" (UniqueName: \"kubernetes.io/projected/6ae7a9e2-889e-499b-8f64-d38976e8d429-kube-api-access-jfznh\") on node \"crc\" DevicePath \"\"" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.619289 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.619297 4838 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7a9e2-889e-499b-8f64-d38976e8d429-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.944780 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" event={"ID":"6ae7a9e2-889e-499b-8f64-d38976e8d429","Type":"ContainerDied","Data":"222cc699676fee880c44797b8f3cb1684a4cdce0789594cb435c7f0bc3c0b104"} Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.944834 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="222cc699676fee880c44797b8f3cb1684a4cdce0789594cb435c7f0bc3c0b104" Dec 07 09:52:29 crc kubenswrapper[4838]: I1207 09:52:29.944880 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-84r6c" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.040349 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng"] Dec 07 09:52:30 crc kubenswrapper[4838]: E1207 09:52:30.040763 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae7a9e2-889e-499b-8f64-d38976e8d429" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.040785 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae7a9e2-889e-499b-8f64-d38976e8d429" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.041060 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae7a9e2-889e-499b-8f64-d38976e8d429" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.041764 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.044905 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.044953 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.046972 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.047114 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.047282 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.047405 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.047512 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.057395 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng"] Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.228933 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.228986 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.229007 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.229061 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.229114 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.229161 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj94j\" (UniqueName: \"kubernetes.io/projected/63630234-9950-43db-a3da-a4ea9ba22be3-kube-api-access-dj94j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.229228 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.331541 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.331836 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.331976 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.332097 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.332209 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.332336 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj94j\" (UniqueName: \"kubernetes.io/projected/63630234-9950-43db-a3da-a4ea9ba22be3-kube-api-access-dj94j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.332450 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.341865 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.346725 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.347225 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.347394 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.347465 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.347857 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.349331 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj94j\" (UniqueName: \"kubernetes.io/projected/63630234-9950-43db-a3da-a4ea9ba22be3-kube-api-access-dj94j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.361041 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.991965 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng"] Dec 07 09:52:30 crc kubenswrapper[4838]: I1207 09:52:30.993028 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:52:31 crc kubenswrapper[4838]: I1207 09:52:31.973157 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" event={"ID":"63630234-9950-43db-a3da-a4ea9ba22be3","Type":"ContainerStarted","Data":"72c8e30578440e894e9bc36e7c491d067f2db3740962cf62d7ea169d60bd9215"} Dec 07 09:52:31 crc kubenswrapper[4838]: I1207 09:52:31.973586 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" event={"ID":"63630234-9950-43db-a3da-a4ea9ba22be3","Type":"ContainerStarted","Data":"8568d85549201fc347cc9cee8cb2fa7baba7405ed6a7aa7d74275d5d23f3f722"} Dec 07 09:52:31 crc kubenswrapper[4838]: I1207 09:52:31.999837 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" podStartSLOduration=1.506494333 podStartE2EDuration="1.999806153s" podCreationTimestamp="2025-12-07 09:52:30 +0000 UTC" firstStartedPulling="2025-12-07 09:52:30.992649362 +0000 UTC m=+2767.699968389" lastFinishedPulling="2025-12-07 09:52:31.485961172 +0000 UTC m=+2768.193280209" observedRunningTime="2025-12-07 09:52:31.995680544 +0000 UTC m=+2768.702999551" watchObservedRunningTime="2025-12-07 09:52:31.999806153 +0000 UTC m=+2768.707125170" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.485483 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2mv2x"] Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.489362 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.496506 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2mv2x"] Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.570373 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-utilities\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.570894 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-catalog-content\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.570951 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxgxz\" (UniqueName: \"kubernetes.io/projected/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-kube-api-access-dxgxz\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.672383 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-catalog-content\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.672546 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxgxz\" (UniqueName: \"kubernetes.io/projected/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-kube-api-access-dxgxz\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.672699 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-utilities\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.673401 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-utilities\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.673401 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-catalog-content\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.692330 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxgxz\" (UniqueName: \"kubernetes.io/projected/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-kube-api-access-dxgxz\") pod \"certified-operators-2mv2x\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:00 crc kubenswrapper[4838]: I1207 09:53:00.810747 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:01 crc kubenswrapper[4838]: I1207 09:53:01.376436 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2mv2x"] Dec 07 09:53:02 crc kubenswrapper[4838]: I1207 09:53:02.270709 4838 generic.go:334] "Generic (PLEG): container finished" podID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerID="8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a" exitCode=0 Dec 07 09:53:02 crc kubenswrapper[4838]: I1207 09:53:02.270893 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mv2x" event={"ID":"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d","Type":"ContainerDied","Data":"8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a"} Dec 07 09:53:02 crc kubenswrapper[4838]: I1207 09:53:02.271081 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mv2x" event={"ID":"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d","Type":"ContainerStarted","Data":"ba2b5463370a32bb407f1bc086a9e1479347cfb9cdb3a4179ef6373dcfbf3426"} Dec 07 09:53:03 crc kubenswrapper[4838]: I1207 09:53:03.286695 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mv2x" event={"ID":"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d","Type":"ContainerStarted","Data":"043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8"} Dec 07 09:53:04 crc kubenswrapper[4838]: I1207 09:53:04.303357 4838 generic.go:334] "Generic (PLEG): container finished" podID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerID="043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8" exitCode=0 Dec 07 09:53:04 crc kubenswrapper[4838]: I1207 09:53:04.303482 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mv2x" event={"ID":"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d","Type":"ContainerDied","Data":"043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8"} Dec 07 09:53:05 crc kubenswrapper[4838]: I1207 09:53:05.314057 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mv2x" event={"ID":"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d","Type":"ContainerStarted","Data":"6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335"} Dec 07 09:53:05 crc kubenswrapper[4838]: I1207 09:53:05.340196 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2mv2x" podStartSLOduration=2.541896459 podStartE2EDuration="5.34017749s" podCreationTimestamp="2025-12-07 09:53:00 +0000 UTC" firstStartedPulling="2025-12-07 09:53:02.272870194 +0000 UTC m=+2798.980189231" lastFinishedPulling="2025-12-07 09:53:05.071151235 +0000 UTC m=+2801.778470262" observedRunningTime="2025-12-07 09:53:05.334506747 +0000 UTC m=+2802.041825764" watchObservedRunningTime="2025-12-07 09:53:05.34017749 +0000 UTC m=+2802.047496507" Dec 07 09:53:10 crc kubenswrapper[4838]: I1207 09:53:10.811125 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:10 crc kubenswrapper[4838]: I1207 09:53:10.812629 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:10 crc kubenswrapper[4838]: I1207 09:53:10.867444 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:11 crc kubenswrapper[4838]: I1207 09:53:11.465417 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:11 crc kubenswrapper[4838]: I1207 09:53:11.523136 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2mv2x"] Dec 07 09:53:13 crc kubenswrapper[4838]: I1207 09:53:13.389573 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2mv2x" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="registry-server" containerID="cri-o://6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335" gracePeriod=2 Dec 07 09:53:13 crc kubenswrapper[4838]: I1207 09:53:13.848993 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.031531 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxgxz\" (UniqueName: \"kubernetes.io/projected/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-kube-api-access-dxgxz\") pod \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.031624 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-catalog-content\") pod \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.031850 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-utilities\") pod \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\" (UID: \"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d\") " Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.032472 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-utilities" (OuterVolumeSpecName: "utilities") pod "e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" (UID: "e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.042036 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-kube-api-access-dxgxz" (OuterVolumeSpecName: "kube-api-access-dxgxz") pod "e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" (UID: "e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d"). InnerVolumeSpecName "kube-api-access-dxgxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.079771 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" (UID: "e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.134367 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.134412 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxgxz\" (UniqueName: \"kubernetes.io/projected/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-kube-api-access-dxgxz\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.134432 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.403671 4838 generic.go:334] "Generic (PLEG): container finished" podID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerID="6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335" exitCode=0 Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.403715 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mv2x" event={"ID":"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d","Type":"ContainerDied","Data":"6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335"} Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.403748 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mv2x" event={"ID":"e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d","Type":"ContainerDied","Data":"ba2b5463370a32bb407f1bc086a9e1479347cfb9cdb3a4179ef6373dcfbf3426"} Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.403768 4838 scope.go:117] "RemoveContainer" containerID="6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.403812 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mv2x" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.431426 4838 scope.go:117] "RemoveContainer" containerID="043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.487371 4838 scope.go:117] "RemoveContainer" containerID="8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.489322 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2mv2x"] Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.506236 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2mv2x"] Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.522052 4838 scope.go:117] "RemoveContainer" containerID="6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335" Dec 07 09:53:14 crc kubenswrapper[4838]: E1207 09:53:14.522566 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335\": container with ID starting with 6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335 not found: ID does not exist" containerID="6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.522605 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335"} err="failed to get container status \"6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335\": rpc error: code = NotFound desc = could not find container \"6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335\": container with ID starting with 6e8e2e12af9a99992eaff92376f36b30502fe17f7fe13fe46024080eb0f41335 not found: ID does not exist" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.522633 4838 scope.go:117] "RemoveContainer" containerID="043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8" Dec 07 09:53:14 crc kubenswrapper[4838]: E1207 09:53:14.522904 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8\": container with ID starting with 043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8 not found: ID does not exist" containerID="043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.522937 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8"} err="failed to get container status \"043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8\": rpc error: code = NotFound desc = could not find container \"043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8\": container with ID starting with 043aff41ab90210675dc37520e3fe08d9111d0e5bf0b6e027d4712c131b1edd8 not found: ID does not exist" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.522959 4838 scope.go:117] "RemoveContainer" containerID="8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a" Dec 07 09:53:14 crc kubenswrapper[4838]: E1207 09:53:14.523189 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a\": container with ID starting with 8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a not found: ID does not exist" containerID="8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a" Dec 07 09:53:14 crc kubenswrapper[4838]: I1207 09:53:14.523215 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a"} err="failed to get container status \"8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a\": rpc error: code = NotFound desc = could not find container \"8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a\": container with ID starting with 8bfada6d28308b42b6bc0e67de2b89e897c5feb8a83480255f200c661616a41a not found: ID does not exist" Dec 07 09:53:15 crc kubenswrapper[4838]: I1207 09:53:15.630764 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" path="/var/lib/kubelet/pods/e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d/volumes" Dec 07 09:53:24 crc kubenswrapper[4838]: I1207 09:53:24.492799 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:53:24 crc kubenswrapper[4838]: I1207 09:53:24.493440 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:53:38 crc kubenswrapper[4838]: I1207 09:53:38.676260 4838 generic.go:334] "Generic (PLEG): container finished" podID="63630234-9950-43db-a3da-a4ea9ba22be3" containerID="72c8e30578440e894e9bc36e7c491d067f2db3740962cf62d7ea169d60bd9215" exitCode=0 Dec 07 09:53:38 crc kubenswrapper[4838]: I1207 09:53:38.676316 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" event={"ID":"63630234-9950-43db-a3da-a4ea9ba22be3","Type":"ContainerDied","Data":"72c8e30578440e894e9bc36e7c491d067f2db3740962cf62d7ea169d60bd9215"} Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.098793 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.264945 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj94j\" (UniqueName: \"kubernetes.io/projected/63630234-9950-43db-a3da-a4ea9ba22be3-kube-api-access-dj94j\") pod \"63630234-9950-43db-a3da-a4ea9ba22be3\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.290503 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-nova-metadata-neutron-config-0\") pod \"63630234-9950-43db-a3da-a4ea9ba22be3\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.290664 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-metadata-combined-ca-bundle\") pod \"63630234-9950-43db-a3da-a4ea9ba22be3\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.290915 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"63630234-9950-43db-a3da-a4ea9ba22be3\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.290946 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-inventory\") pod \"63630234-9950-43db-a3da-a4ea9ba22be3\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.291259 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ceph\") pod \"63630234-9950-43db-a3da-a4ea9ba22be3\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.291309 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ssh-key\") pod \"63630234-9950-43db-a3da-a4ea9ba22be3\" (UID: \"63630234-9950-43db-a3da-a4ea9ba22be3\") " Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.334984 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63630234-9950-43db-a3da-a4ea9ba22be3-kube-api-access-dj94j" (OuterVolumeSpecName: "kube-api-access-dj94j") pod "63630234-9950-43db-a3da-a4ea9ba22be3" (UID: "63630234-9950-43db-a3da-a4ea9ba22be3"). InnerVolumeSpecName "kube-api-access-dj94j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.335171 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "63630234-9950-43db-a3da-a4ea9ba22be3" (UID: "63630234-9950-43db-a3da-a4ea9ba22be3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.334880 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ceph" (OuterVolumeSpecName: "ceph") pod "63630234-9950-43db-a3da-a4ea9ba22be3" (UID: "63630234-9950-43db-a3da-a4ea9ba22be3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.384978 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "63630234-9950-43db-a3da-a4ea9ba22be3" (UID: "63630234-9950-43db-a3da-a4ea9ba22be3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.393946 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-inventory" (OuterVolumeSpecName: "inventory") pod "63630234-9950-43db-a3da-a4ea9ba22be3" (UID: "63630234-9950-43db-a3da-a4ea9ba22be3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.398002 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj94j\" (UniqueName: \"kubernetes.io/projected/63630234-9950-43db-a3da-a4ea9ba22be3-kube-api-access-dj94j\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.398033 4838 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.398046 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.398056 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.398065 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.418073 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "63630234-9950-43db-a3da-a4ea9ba22be3" (UID: "63630234-9950-43db-a3da-a4ea9ba22be3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.435906 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "63630234-9950-43db-a3da-a4ea9ba22be3" (UID: "63630234-9950-43db-a3da-a4ea9ba22be3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.501579 4838 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.501634 4838 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/63630234-9950-43db-a3da-a4ea9ba22be3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.697478 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" event={"ID":"63630234-9950-43db-a3da-a4ea9ba22be3","Type":"ContainerDied","Data":"8568d85549201fc347cc9cee8cb2fa7baba7405ed6a7aa7d74275d5d23f3f722"} Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.697522 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8568d85549201fc347cc9cee8cb2fa7baba7405ed6a7aa7d74275d5d23f3f722" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.697542 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.863472 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt"] Dec 07 09:53:40 crc kubenswrapper[4838]: E1207 09:53:40.863918 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="registry-server" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.863935 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="registry-server" Dec 07 09:53:40 crc kubenswrapper[4838]: E1207 09:53:40.863958 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63630234-9950-43db-a3da-a4ea9ba22be3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.863968 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="63630234-9950-43db-a3da-a4ea9ba22be3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 07 09:53:40 crc kubenswrapper[4838]: E1207 09:53:40.863986 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="extract-content" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.863992 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="extract-content" Dec 07 09:53:40 crc kubenswrapper[4838]: E1207 09:53:40.864006 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="extract-utilities" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.864013 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="extract-utilities" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.864215 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10da7b3-6fe7-4f08-8d86-055b2ceb6c2d" containerName="registry-server" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.864244 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="63630234-9950-43db-a3da-a4ea9ba22be3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.865072 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.868338 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.868409 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.868431 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.868617 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.869804 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.870025 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.870680 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt"] Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.910986 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.911042 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.911097 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.911170 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.911237 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sxv5\" (UniqueName: \"kubernetes.io/projected/47331703-36e3-47e2-bbe0-e891130bdbe7-kube-api-access-8sxv5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:40 crc kubenswrapper[4838]: I1207 09:53:40.911255 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.013501 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.013593 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.013663 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.013773 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.013907 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sxv5\" (UniqueName: \"kubernetes.io/projected/47331703-36e3-47e2-bbe0-e891130bdbe7-kube-api-access-8sxv5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.013943 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.019673 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.019706 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.020755 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.021325 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.023758 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.046471 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sxv5\" (UniqueName: \"kubernetes.io/projected/47331703-36e3-47e2-bbe0-e891130bdbe7-kube-api-access-8sxv5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.182234 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.551339 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt"] Dec 07 09:53:41 crc kubenswrapper[4838]: I1207 09:53:41.707560 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" event={"ID":"47331703-36e3-47e2-bbe0-e891130bdbe7","Type":"ContainerStarted","Data":"65bbdcf67576fcc8bfc5ec33e5e524f3eff26ec25213f791e050d57200b6650d"} Dec 07 09:53:42 crc kubenswrapper[4838]: I1207 09:53:42.717024 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" event={"ID":"47331703-36e3-47e2-bbe0-e891130bdbe7","Type":"ContainerStarted","Data":"67779eae3c88df8d1b85857aee66edfeed6bc79173e0f52653071fab07b4291a"} Dec 07 09:53:42 crc kubenswrapper[4838]: I1207 09:53:42.742109 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" podStartSLOduration=2.326882157 podStartE2EDuration="2.742088696s" podCreationTimestamp="2025-12-07 09:53:40 +0000 UTC" firstStartedPulling="2025-12-07 09:53:41.558961611 +0000 UTC m=+2838.266280628" lastFinishedPulling="2025-12-07 09:53:41.97416811 +0000 UTC m=+2838.681487167" observedRunningTime="2025-12-07 09:53:42.741051316 +0000 UTC m=+2839.448370353" watchObservedRunningTime="2025-12-07 09:53:42.742088696 +0000 UTC m=+2839.449407723" Dec 07 09:53:54 crc kubenswrapper[4838]: I1207 09:53:54.492983 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:53:54 crc kubenswrapper[4838]: I1207 09:53:54.493688 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.791142 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v5h6c"] Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.817932 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.849810 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5h6c"] Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.867559 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fkhf\" (UniqueName: \"kubernetes.io/projected/20955b73-8474-4969-802c-878d2fd9b293-kube-api-access-5fkhf\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.867674 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-utilities\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.867707 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-catalog-content\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.968331 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-catalog-content\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.968403 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fkhf\" (UniqueName: \"kubernetes.io/projected/20955b73-8474-4969-802c-878d2fd9b293-kube-api-access-5fkhf\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.968499 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-utilities\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.968927 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-catalog-content\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.969145 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-utilities\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:58 crc kubenswrapper[4838]: I1207 09:53:58.988131 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fkhf\" (UniqueName: \"kubernetes.io/projected/20955b73-8474-4969-802c-878d2fd9b293-kube-api-access-5fkhf\") pod \"community-operators-v5h6c\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:59 crc kubenswrapper[4838]: I1207 09:53:59.162661 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:53:59 crc kubenswrapper[4838]: W1207 09:53:59.718773 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20955b73_8474_4969_802c_878d2fd9b293.slice/crio-1d7d66b75c32ff6bdbbe5d5c810d4bce038a4fd9bbf2a0d859d297080ce89323 WatchSource:0}: Error finding container 1d7d66b75c32ff6bdbbe5d5c810d4bce038a4fd9bbf2a0d859d297080ce89323: Status 404 returned error can't find the container with id 1d7d66b75c32ff6bdbbe5d5c810d4bce038a4fd9bbf2a0d859d297080ce89323 Dec 07 09:53:59 crc kubenswrapper[4838]: I1207 09:53:59.719580 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5h6c"] Dec 07 09:53:59 crc kubenswrapper[4838]: I1207 09:53:59.880831 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5h6c" event={"ID":"20955b73-8474-4969-802c-878d2fd9b293","Type":"ContainerStarted","Data":"1d7d66b75c32ff6bdbbe5d5c810d4bce038a4fd9bbf2a0d859d297080ce89323"} Dec 07 09:54:00 crc kubenswrapper[4838]: I1207 09:54:00.889553 4838 generic.go:334] "Generic (PLEG): container finished" podID="20955b73-8474-4969-802c-878d2fd9b293" containerID="20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813" exitCode=0 Dec 07 09:54:00 crc kubenswrapper[4838]: I1207 09:54:00.889620 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5h6c" event={"ID":"20955b73-8474-4969-802c-878d2fd9b293","Type":"ContainerDied","Data":"20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813"} Dec 07 09:54:01 crc kubenswrapper[4838]: I1207 09:54:01.901314 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5h6c" event={"ID":"20955b73-8474-4969-802c-878d2fd9b293","Type":"ContainerStarted","Data":"e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58"} Dec 07 09:54:03 crc kubenswrapper[4838]: I1207 09:54:03.938998 4838 generic.go:334] "Generic (PLEG): container finished" podID="20955b73-8474-4969-802c-878d2fd9b293" containerID="e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58" exitCode=0 Dec 07 09:54:03 crc kubenswrapper[4838]: I1207 09:54:03.939073 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5h6c" event={"ID":"20955b73-8474-4969-802c-878d2fd9b293","Type":"ContainerDied","Data":"e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58"} Dec 07 09:54:04 crc kubenswrapper[4838]: I1207 09:54:04.951546 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5h6c" event={"ID":"20955b73-8474-4969-802c-878d2fd9b293","Type":"ContainerStarted","Data":"bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271"} Dec 07 09:54:04 crc kubenswrapper[4838]: I1207 09:54:04.982740 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v5h6c" podStartSLOduration=3.525071811 podStartE2EDuration="6.982717627s" podCreationTimestamp="2025-12-07 09:53:58 +0000 UTC" firstStartedPulling="2025-12-07 09:54:00.893717085 +0000 UTC m=+2857.601036102" lastFinishedPulling="2025-12-07 09:54:04.351362891 +0000 UTC m=+2861.058681918" observedRunningTime="2025-12-07 09:54:04.974055073 +0000 UTC m=+2861.681374140" watchObservedRunningTime="2025-12-07 09:54:04.982717627 +0000 UTC m=+2861.690036654" Dec 07 09:54:09 crc kubenswrapper[4838]: I1207 09:54:09.163291 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:54:09 crc kubenswrapper[4838]: I1207 09:54:09.163938 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:54:09 crc kubenswrapper[4838]: I1207 09:54:09.243486 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:54:10 crc kubenswrapper[4838]: I1207 09:54:10.030362 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:54:10 crc kubenswrapper[4838]: I1207 09:54:10.766117 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5h6c"] Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.008094 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v5h6c" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="registry-server" containerID="cri-o://bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271" gracePeriod=2 Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.465496 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.526483 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-utilities\") pod \"20955b73-8474-4969-802c-878d2fd9b293\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.526581 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-catalog-content\") pod \"20955b73-8474-4969-802c-878d2fd9b293\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.526690 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fkhf\" (UniqueName: \"kubernetes.io/projected/20955b73-8474-4969-802c-878d2fd9b293-kube-api-access-5fkhf\") pod \"20955b73-8474-4969-802c-878d2fd9b293\" (UID: \"20955b73-8474-4969-802c-878d2fd9b293\") " Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.527467 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-utilities" (OuterVolumeSpecName: "utilities") pod "20955b73-8474-4969-802c-878d2fd9b293" (UID: "20955b73-8474-4969-802c-878d2fd9b293"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.539540 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20955b73-8474-4969-802c-878d2fd9b293-kube-api-access-5fkhf" (OuterVolumeSpecName: "kube-api-access-5fkhf") pod "20955b73-8474-4969-802c-878d2fd9b293" (UID: "20955b73-8474-4969-802c-878d2fd9b293"). InnerVolumeSpecName "kube-api-access-5fkhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.575359 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20955b73-8474-4969-802c-878d2fd9b293" (UID: "20955b73-8474-4969-802c-878d2fd9b293"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.627986 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.628012 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20955b73-8474-4969-802c-878d2fd9b293-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 09:54:12 crc kubenswrapper[4838]: I1207 09:54:12.628022 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fkhf\" (UniqueName: \"kubernetes.io/projected/20955b73-8474-4969-802c-878d2fd9b293-kube-api-access-5fkhf\") on node \"crc\" DevicePath \"\"" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.017195 4838 generic.go:334] "Generic (PLEG): container finished" podID="20955b73-8474-4969-802c-878d2fd9b293" containerID="bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271" exitCode=0 Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.017251 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5h6c" event={"ID":"20955b73-8474-4969-802c-878d2fd9b293","Type":"ContainerDied","Data":"bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271"} Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.017298 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5h6c" event={"ID":"20955b73-8474-4969-802c-878d2fd9b293","Type":"ContainerDied","Data":"1d7d66b75c32ff6bdbbe5d5c810d4bce038a4fd9bbf2a0d859d297080ce89323"} Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.017349 4838 scope.go:117] "RemoveContainer" containerID="bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.018610 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5h6c" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.047284 4838 scope.go:117] "RemoveContainer" containerID="e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.087332 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5h6c"] Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.104539 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v5h6c"] Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.116954 4838 scope.go:117] "RemoveContainer" containerID="20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.134610 4838 scope.go:117] "RemoveContainer" containerID="bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271" Dec 07 09:54:13 crc kubenswrapper[4838]: E1207 09:54:13.135127 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271\": container with ID starting with bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271 not found: ID does not exist" containerID="bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.135155 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271"} err="failed to get container status \"bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271\": rpc error: code = NotFound desc = could not find container \"bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271\": container with ID starting with bcc7cb114d3e587d444ebb187856db64310ecaaca3b59e15894821509616f271 not found: ID does not exist" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.135174 4838 scope.go:117] "RemoveContainer" containerID="e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58" Dec 07 09:54:13 crc kubenswrapper[4838]: E1207 09:54:13.135474 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58\": container with ID starting with e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58 not found: ID does not exist" containerID="e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.135496 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58"} err="failed to get container status \"e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58\": rpc error: code = NotFound desc = could not find container \"e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58\": container with ID starting with e5e99adad30cefb9c7c89c845cf07136a9842020495e04a286c14abb28785f58 not found: ID does not exist" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.135510 4838 scope.go:117] "RemoveContainer" containerID="20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813" Dec 07 09:54:13 crc kubenswrapper[4838]: E1207 09:54:13.135973 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813\": container with ID starting with 20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813 not found: ID does not exist" containerID="20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.136048 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813"} err="failed to get container status \"20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813\": rpc error: code = NotFound desc = could not find container \"20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813\": container with ID starting with 20e586b04871397bc15ccd959ee778c7c02f4a565e1ad9819f4ee5d1ad1a5813 not found: ID does not exist" Dec 07 09:54:13 crc kubenswrapper[4838]: I1207 09:54:13.624378 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20955b73-8474-4969-802c-878d2fd9b293" path="/var/lib/kubelet/pods/20955b73-8474-4969-802c-878d2fd9b293/volumes" Dec 07 09:54:24 crc kubenswrapper[4838]: I1207 09:54:24.492933 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 09:54:24 crc kubenswrapper[4838]: I1207 09:54:24.493641 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 09:54:24 crc kubenswrapper[4838]: I1207 09:54:24.493697 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 09:54:24 crc kubenswrapper[4838]: I1207 09:54:24.494532 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 09:54:24 crc kubenswrapper[4838]: I1207 09:54:24.494607 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" gracePeriod=600 Dec 07 09:54:24 crc kubenswrapper[4838]: E1207 09:54:24.618141 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:54:25 crc kubenswrapper[4838]: I1207 09:54:25.133576 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" exitCode=0 Dec 07 09:54:25 crc kubenswrapper[4838]: I1207 09:54:25.133678 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f"} Dec 07 09:54:25 crc kubenswrapper[4838]: I1207 09:54:25.133977 4838 scope.go:117] "RemoveContainer" containerID="a37d598a4ad1f8e326405a22ca30de8c4dc862ce75ea17fb6791106fb71d0058" Dec 07 09:54:25 crc kubenswrapper[4838]: I1207 09:54:25.134681 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:54:25 crc kubenswrapper[4838]: E1207 09:54:25.134949 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:54:40 crc kubenswrapper[4838]: I1207 09:54:40.614511 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:54:40 crc kubenswrapper[4838]: E1207 09:54:40.615443 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:54:51 crc kubenswrapper[4838]: I1207 09:54:51.614897 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:54:51 crc kubenswrapper[4838]: E1207 09:54:51.615844 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:55:05 crc kubenswrapper[4838]: I1207 09:55:05.615139 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:55:05 crc kubenswrapper[4838]: E1207 09:55:05.616535 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:55:20 crc kubenswrapper[4838]: I1207 09:55:20.614290 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:55:20 crc kubenswrapper[4838]: E1207 09:55:20.615023 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:55:34 crc kubenswrapper[4838]: I1207 09:55:34.614469 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:55:34 crc kubenswrapper[4838]: E1207 09:55:34.615313 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:55:46 crc kubenswrapper[4838]: I1207 09:55:46.615128 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:55:46 crc kubenswrapper[4838]: E1207 09:55:46.615999 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:56:00 crc kubenswrapper[4838]: I1207 09:56:00.617000 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:56:00 crc kubenswrapper[4838]: E1207 09:56:00.619372 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:56:13 crc kubenswrapper[4838]: I1207 09:56:13.621300 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:56:13 crc kubenswrapper[4838]: E1207 09:56:13.622110 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:56:26 crc kubenswrapper[4838]: I1207 09:56:26.615369 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:56:26 crc kubenswrapper[4838]: E1207 09:56:26.616268 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:56:38 crc kubenswrapper[4838]: I1207 09:56:38.615597 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:56:38 crc kubenswrapper[4838]: E1207 09:56:38.616808 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:56:49 crc kubenswrapper[4838]: I1207 09:56:49.615006 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:56:49 crc kubenswrapper[4838]: E1207 09:56:49.616557 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:57:04 crc kubenswrapper[4838]: I1207 09:57:04.614759 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:57:04 crc kubenswrapper[4838]: E1207 09:57:04.615567 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:57:19 crc kubenswrapper[4838]: I1207 09:57:19.614805 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:57:19 crc kubenswrapper[4838]: E1207 09:57:19.615922 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:57:31 crc kubenswrapper[4838]: I1207 09:57:31.615202 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:57:31 crc kubenswrapper[4838]: E1207 09:57:31.616231 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:57:44 crc kubenswrapper[4838]: I1207 09:57:44.614170 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:57:44 crc kubenswrapper[4838]: E1207 09:57:44.614867 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:57:59 crc kubenswrapper[4838]: I1207 09:57:59.616078 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:57:59 crc kubenswrapper[4838]: E1207 09:57:59.616761 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:58:14 crc kubenswrapper[4838]: I1207 09:58:14.614525 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:58:14 crc kubenswrapper[4838]: E1207 09:58:14.615377 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:58:27 crc kubenswrapper[4838]: I1207 09:58:27.614570 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:58:27 crc kubenswrapper[4838]: E1207 09:58:27.616361 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:58:33 crc kubenswrapper[4838]: I1207 09:58:33.725737 4838 generic.go:334] "Generic (PLEG): container finished" podID="47331703-36e3-47e2-bbe0-e891130bdbe7" containerID="67779eae3c88df8d1b85857aee66edfeed6bc79173e0f52653071fab07b4291a" exitCode=0 Dec 07 09:58:33 crc kubenswrapper[4838]: I1207 09:58:33.725802 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" event={"ID":"47331703-36e3-47e2-bbe0-e891130bdbe7","Type":"ContainerDied","Data":"67779eae3c88df8d1b85857aee66edfeed6bc79173e0f52653071fab07b4291a"} Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.199376 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.387055 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sxv5\" (UniqueName: \"kubernetes.io/projected/47331703-36e3-47e2-bbe0-e891130bdbe7-kube-api-access-8sxv5\") pod \"47331703-36e3-47e2-bbe0-e891130bdbe7\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.387152 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ceph\") pod \"47331703-36e3-47e2-bbe0-e891130bdbe7\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.387234 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-inventory\") pod \"47331703-36e3-47e2-bbe0-e891130bdbe7\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.387298 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-secret-0\") pod \"47331703-36e3-47e2-bbe0-e891130bdbe7\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.387429 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ssh-key\") pod \"47331703-36e3-47e2-bbe0-e891130bdbe7\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.387455 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-combined-ca-bundle\") pod \"47331703-36e3-47e2-bbe0-e891130bdbe7\" (UID: \"47331703-36e3-47e2-bbe0-e891130bdbe7\") " Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.394690 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47331703-36e3-47e2-bbe0-e891130bdbe7-kube-api-access-8sxv5" (OuterVolumeSpecName: "kube-api-access-8sxv5") pod "47331703-36e3-47e2-bbe0-e891130bdbe7" (UID: "47331703-36e3-47e2-bbe0-e891130bdbe7"). InnerVolumeSpecName "kube-api-access-8sxv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.397036 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "47331703-36e3-47e2-bbe0-e891130bdbe7" (UID: "47331703-36e3-47e2-bbe0-e891130bdbe7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.397249 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ceph" (OuterVolumeSpecName: "ceph") pod "47331703-36e3-47e2-bbe0-e891130bdbe7" (UID: "47331703-36e3-47e2-bbe0-e891130bdbe7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.419920 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "47331703-36e3-47e2-bbe0-e891130bdbe7" (UID: "47331703-36e3-47e2-bbe0-e891130bdbe7"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.424189 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-inventory" (OuterVolumeSpecName: "inventory") pod "47331703-36e3-47e2-bbe0-e891130bdbe7" (UID: "47331703-36e3-47e2-bbe0-e891130bdbe7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.429468 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "47331703-36e3-47e2-bbe0-e891130bdbe7" (UID: "47331703-36e3-47e2-bbe0-e891130bdbe7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.489831 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.489855 4838 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.489889 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sxv5\" (UniqueName: \"kubernetes.io/projected/47331703-36e3-47e2-bbe0-e891130bdbe7-kube-api-access-8sxv5\") on node \"crc\" DevicePath \"\"" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.489901 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.489909 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.489920 4838 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/47331703-36e3-47e2-bbe0-e891130bdbe7-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.752774 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" event={"ID":"47331703-36e3-47e2-bbe0-e891130bdbe7","Type":"ContainerDied","Data":"65bbdcf67576fcc8bfc5ec33e5e524f3eff26ec25213f791e050d57200b6650d"} Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.752861 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65bbdcf67576fcc8bfc5ec33e5e524f3eff26ec25213f791e050d57200b6650d" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.753164 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.863156 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd"] Dec 07 09:58:35 crc kubenswrapper[4838]: E1207 09:58:35.863566 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="registry-server" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.863586 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="registry-server" Dec 07 09:58:35 crc kubenswrapper[4838]: E1207 09:58:35.863605 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="extract-content" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.863613 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="extract-content" Dec 07 09:58:35 crc kubenswrapper[4838]: E1207 09:58:35.863629 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="extract-utilities" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.863638 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="extract-utilities" Dec 07 09:58:35 crc kubenswrapper[4838]: E1207 09:58:35.863653 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47331703-36e3-47e2-bbe0-e891130bdbe7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.863663 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="47331703-36e3-47e2-bbe0-e891130bdbe7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.863912 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="47331703-36e3-47e2-bbe0-e891130bdbe7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.863944 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="20955b73-8474-4969-802c-878d2fd9b293" containerName="registry-server" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.864617 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.870520 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.870617 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.871033 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.871102 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.871393 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.871048 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.876334 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.871109 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.876608 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sntg6" Dec 07 09:58:35 crc kubenswrapper[4838]: I1207 09:58:35.880439 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd"] Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004235 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004338 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004384 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004423 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004456 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004486 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004530 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004612 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hjq9\" (UniqueName: \"kubernetes.io/projected/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-kube-api-access-4hjq9\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004640 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004674 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.004709 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.106380 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107163 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107201 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107226 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107245 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107275 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107315 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hjq9\" (UniqueName: \"kubernetes.io/projected/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-kube-api-access-4hjq9\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107337 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107362 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107385 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.107414 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.108427 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.108501 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.110285 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.113226 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.113247 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.113732 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.116391 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.116494 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.117936 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.119260 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.130682 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hjq9\" (UniqueName: \"kubernetes.io/projected/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-kube-api-access-4hjq9\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.188528 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.778927 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd"] Dec 07 09:58:36 crc kubenswrapper[4838]: I1207 09:58:36.784589 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 09:58:37 crc kubenswrapper[4838]: I1207 09:58:37.771160 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" event={"ID":"e47e7bbd-a623-4a13-ba01-2fb7ff984b40","Type":"ContainerStarted","Data":"978dd07beeb62d54f0961e800bc1d177c4156fa25e9a2c6b5b2e196bedcd646e"} Dec 07 09:58:37 crc kubenswrapper[4838]: I1207 09:58:37.771613 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" event={"ID":"e47e7bbd-a623-4a13-ba01-2fb7ff984b40","Type":"ContainerStarted","Data":"ebe54e3962626f6d7f26a1972051a2995270cc2626575a5ddc63cc5ba3181846"} Dec 07 09:58:37 crc kubenswrapper[4838]: I1207 09:58:37.798005 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" podStartSLOduration=2.370714836 podStartE2EDuration="2.797982968s" podCreationTimestamp="2025-12-07 09:58:35 +0000 UTC" firstStartedPulling="2025-12-07 09:58:36.784327554 +0000 UTC m=+3133.491646591" lastFinishedPulling="2025-12-07 09:58:37.211595696 +0000 UTC m=+3133.918914723" observedRunningTime="2025-12-07 09:58:37.787244794 +0000 UTC m=+3134.494563811" watchObservedRunningTime="2025-12-07 09:58:37.797982968 +0000 UTC m=+3134.505301985" Dec 07 09:58:39 crc kubenswrapper[4838]: I1207 09:58:39.616242 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:58:39 crc kubenswrapper[4838]: E1207 09:58:39.617128 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:58:54 crc kubenswrapper[4838]: I1207 09:58:54.614922 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:58:54 crc kubenswrapper[4838]: E1207 09:58:54.615706 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:59:08 crc kubenswrapper[4838]: I1207 09:59:08.614176 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:59:08 crc kubenswrapper[4838]: E1207 09:59:08.614842 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:59:22 crc kubenswrapper[4838]: I1207 09:59:22.614017 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:59:22 crc kubenswrapper[4838]: E1207 09:59:22.614759 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 09:59:37 crc kubenswrapper[4838]: I1207 09:59:37.614829 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 09:59:38 crc kubenswrapper[4838]: I1207 09:59:38.444473 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"9f7e06fe287c32a6d89d0c626a75199a006e5e2536c384a448e44b3eb61037c0"} Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.154324 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq"] Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.156088 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.159719 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.161646 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.169868 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq"] Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.211284 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4197083e-4822-46b9-acca-8fae27debdad-secret-volume\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.211328 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4197083e-4822-46b9-acca-8fae27debdad-config-volume\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.211455 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llnbw\" (UniqueName: \"kubernetes.io/projected/4197083e-4822-46b9-acca-8fae27debdad-kube-api-access-llnbw\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.313092 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llnbw\" (UniqueName: \"kubernetes.io/projected/4197083e-4822-46b9-acca-8fae27debdad-kube-api-access-llnbw\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.313186 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4197083e-4822-46b9-acca-8fae27debdad-secret-volume\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.313224 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4197083e-4822-46b9-acca-8fae27debdad-config-volume\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.314457 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4197083e-4822-46b9-acca-8fae27debdad-config-volume\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.320208 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4197083e-4822-46b9-acca-8fae27debdad-secret-volume\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.329027 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llnbw\" (UniqueName: \"kubernetes.io/projected/4197083e-4822-46b9-acca-8fae27debdad-kube-api-access-llnbw\") pod \"collect-profiles-29418360-hljsq\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.509630 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:00 crc kubenswrapper[4838]: I1207 10:00:00.995129 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq"] Dec 07 10:00:01 crc kubenswrapper[4838]: I1207 10:00:01.825449 4838 generic.go:334] "Generic (PLEG): container finished" podID="4197083e-4822-46b9-acca-8fae27debdad" containerID="1df9cbae5988125b0032f07560044d451bd69c7720cb04a2aedab4ce5f97cc4a" exitCode=0 Dec 07 10:00:01 crc kubenswrapper[4838]: I1207 10:00:01.825500 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" event={"ID":"4197083e-4822-46b9-acca-8fae27debdad","Type":"ContainerDied","Data":"1df9cbae5988125b0032f07560044d451bd69c7720cb04a2aedab4ce5f97cc4a"} Dec 07 10:00:01 crc kubenswrapper[4838]: I1207 10:00:01.825541 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" event={"ID":"4197083e-4822-46b9-acca-8fae27debdad","Type":"ContainerStarted","Data":"37ab0fa4d46bab20a25a21a82c27ac6c528c10de82e7da349a482917fbce04dc"} Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.206458 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.265080 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4197083e-4822-46b9-acca-8fae27debdad-config-volume\") pod \"4197083e-4822-46b9-acca-8fae27debdad\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.265485 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4197083e-4822-46b9-acca-8fae27debdad-secret-volume\") pod \"4197083e-4822-46b9-acca-8fae27debdad\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.265680 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llnbw\" (UniqueName: \"kubernetes.io/projected/4197083e-4822-46b9-acca-8fae27debdad-kube-api-access-llnbw\") pod \"4197083e-4822-46b9-acca-8fae27debdad\" (UID: \"4197083e-4822-46b9-acca-8fae27debdad\") " Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.266202 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4197083e-4822-46b9-acca-8fae27debdad-config-volume" (OuterVolumeSpecName: "config-volume") pod "4197083e-4822-46b9-acca-8fae27debdad" (UID: "4197083e-4822-46b9-acca-8fae27debdad"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.267011 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4197083e-4822-46b9-acca-8fae27debdad-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.283054 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4197083e-4822-46b9-acca-8fae27debdad-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4197083e-4822-46b9-acca-8fae27debdad" (UID: "4197083e-4822-46b9-acca-8fae27debdad"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.283115 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4197083e-4822-46b9-acca-8fae27debdad-kube-api-access-llnbw" (OuterVolumeSpecName: "kube-api-access-llnbw") pod "4197083e-4822-46b9-acca-8fae27debdad" (UID: "4197083e-4822-46b9-acca-8fae27debdad"). InnerVolumeSpecName "kube-api-access-llnbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.368477 4838 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4197083e-4822-46b9-acca-8fae27debdad-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.368775 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llnbw\" (UniqueName: \"kubernetes.io/projected/4197083e-4822-46b9-acca-8fae27debdad-kube-api-access-llnbw\") on node \"crc\" DevicePath \"\"" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.847748 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" event={"ID":"4197083e-4822-46b9-acca-8fae27debdad","Type":"ContainerDied","Data":"37ab0fa4d46bab20a25a21a82c27ac6c528c10de82e7da349a482917fbce04dc"} Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.847787 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37ab0fa4d46bab20a25a21a82c27ac6c528c10de82e7da349a482917fbce04dc" Dec 07 10:00:03 crc kubenswrapper[4838]: I1207 10:00:03.847888 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418360-hljsq" Dec 07 10:00:04 crc kubenswrapper[4838]: I1207 10:00:04.301204 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q"] Dec 07 10:00:04 crc kubenswrapper[4838]: I1207 10:00:04.307833 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418315-4ph5q"] Dec 07 10:00:05 crc kubenswrapper[4838]: I1207 10:00:05.631588 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9083d71b-2831-4a45-80b2-baa769ed181f" path="/var/lib/kubelet/pods/9083d71b-2831-4a45-80b2-baa769ed181f/volumes" Dec 07 10:00:53 crc kubenswrapper[4838]: I1207 10:00:53.111286 4838 scope.go:117] "RemoveContainer" containerID="617037b1b5c9e73c95d2534390edb323f8fad7af92a6df7183339bf2463858e9" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.164954 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29418361-mnpmk"] Dec 07 10:01:00 crc kubenswrapper[4838]: E1207 10:01:00.165901 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4197083e-4822-46b9-acca-8fae27debdad" containerName="collect-profiles" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.165918 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="4197083e-4822-46b9-acca-8fae27debdad" containerName="collect-profiles" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.166169 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="4197083e-4822-46b9-acca-8fae27debdad" containerName="collect-profiles" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.168718 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.176044 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29418361-mnpmk"] Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.255738 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-combined-ca-bundle\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.255779 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hvsd\" (UniqueName: \"kubernetes.io/projected/5de7aef2-84f7-4398-a557-4041f9031e31-kube-api-access-2hvsd\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.255811 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-fernet-keys\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.255884 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-config-data\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.357196 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hvsd\" (UniqueName: \"kubernetes.io/projected/5de7aef2-84f7-4398-a557-4041f9031e31-kube-api-access-2hvsd\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.357300 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-fernet-keys\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.357382 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-config-data\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.357464 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-combined-ca-bundle\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.364037 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-config-data\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.364262 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-fernet-keys\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.365570 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-combined-ca-bundle\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.392020 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hvsd\" (UniqueName: \"kubernetes.io/projected/5de7aef2-84f7-4398-a557-4041f9031e31-kube-api-access-2hvsd\") pod \"keystone-cron-29418361-mnpmk\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.489284 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:00 crc kubenswrapper[4838]: I1207 10:01:00.977177 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29418361-mnpmk"] Dec 07 10:01:01 crc kubenswrapper[4838]: I1207 10:01:01.407834 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418361-mnpmk" event={"ID":"5de7aef2-84f7-4398-a557-4041f9031e31","Type":"ContainerStarted","Data":"d47fb06c5f9319ff6c4e9495a42b117d98fe96b76823e85f58d7e4325996be7a"} Dec 07 10:01:01 crc kubenswrapper[4838]: I1207 10:01:01.408199 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418361-mnpmk" event={"ID":"5de7aef2-84f7-4398-a557-4041f9031e31","Type":"ContainerStarted","Data":"b70df23fe1b5c7b29cf69bcc3ec5485d6c5b1e1387b39a9e220dcdada360af1d"} Dec 07 10:01:01 crc kubenswrapper[4838]: I1207 10:01:01.435939 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29418361-mnpmk" podStartSLOduration=1.435920169 podStartE2EDuration="1.435920169s" podCreationTimestamp="2025-12-07 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:01:01.429640143 +0000 UTC m=+3278.136959190" watchObservedRunningTime="2025-12-07 10:01:01.435920169 +0000 UTC m=+3278.143239186" Dec 07 10:01:03 crc kubenswrapper[4838]: I1207 10:01:03.424103 4838 generic.go:334] "Generic (PLEG): container finished" podID="5de7aef2-84f7-4398-a557-4041f9031e31" containerID="d47fb06c5f9319ff6c4e9495a42b117d98fe96b76823e85f58d7e4325996be7a" exitCode=0 Dec 07 10:01:03 crc kubenswrapper[4838]: I1207 10:01:03.424176 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418361-mnpmk" event={"ID":"5de7aef2-84f7-4398-a557-4041f9031e31","Type":"ContainerDied","Data":"d47fb06c5f9319ff6c4e9495a42b117d98fe96b76823e85f58d7e4325996be7a"} Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.741653 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.931704 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hvsd\" (UniqueName: \"kubernetes.io/projected/5de7aef2-84f7-4398-a557-4041f9031e31-kube-api-access-2hvsd\") pod \"5de7aef2-84f7-4398-a557-4041f9031e31\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.932127 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-config-data\") pod \"5de7aef2-84f7-4398-a557-4041f9031e31\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.932345 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-combined-ca-bundle\") pod \"5de7aef2-84f7-4398-a557-4041f9031e31\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.932502 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-fernet-keys\") pod \"5de7aef2-84f7-4398-a557-4041f9031e31\" (UID: \"5de7aef2-84f7-4398-a557-4041f9031e31\") " Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.937638 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5de7aef2-84f7-4398-a557-4041f9031e31" (UID: "5de7aef2-84f7-4398-a557-4041f9031e31"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.949781 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de7aef2-84f7-4398-a557-4041f9031e31-kube-api-access-2hvsd" (OuterVolumeSpecName: "kube-api-access-2hvsd") pod "5de7aef2-84f7-4398-a557-4041f9031e31" (UID: "5de7aef2-84f7-4398-a557-4041f9031e31"). InnerVolumeSpecName "kube-api-access-2hvsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.962918 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5de7aef2-84f7-4398-a557-4041f9031e31" (UID: "5de7aef2-84f7-4398-a557-4041f9031e31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:01:04 crc kubenswrapper[4838]: I1207 10:01:04.981202 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-config-data" (OuterVolumeSpecName: "config-data") pod "5de7aef2-84f7-4398-a557-4041f9031e31" (UID: "5de7aef2-84f7-4398-a557-4041f9031e31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:01:05 crc kubenswrapper[4838]: I1207 10:01:05.035438 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hvsd\" (UniqueName: \"kubernetes.io/projected/5de7aef2-84f7-4398-a557-4041f9031e31-kube-api-access-2hvsd\") on node \"crc\" DevicePath \"\"" Dec 07 10:01:05 crc kubenswrapper[4838]: I1207 10:01:05.035484 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:01:05 crc kubenswrapper[4838]: I1207 10:01:05.035503 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:01:05 crc kubenswrapper[4838]: I1207 10:01:05.035522 4838 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5de7aef2-84f7-4398-a557-4041f9031e31-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 07 10:01:05 crc kubenswrapper[4838]: I1207 10:01:05.447701 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418361-mnpmk" event={"ID":"5de7aef2-84f7-4398-a557-4041f9031e31","Type":"ContainerDied","Data":"b70df23fe1b5c7b29cf69bcc3ec5485d6c5b1e1387b39a9e220dcdada360af1d"} Dec 07 10:01:05 crc kubenswrapper[4838]: I1207 10:01:05.447747 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b70df23fe1b5c7b29cf69bcc3ec5485d6c5b1e1387b39a9e220dcdada360af1d" Dec 07 10:01:05 crc kubenswrapper[4838]: I1207 10:01:05.447812 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418361-mnpmk" Dec 07 10:01:54 crc kubenswrapper[4838]: I1207 10:01:54.493561 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:01:54 crc kubenswrapper[4838]: I1207 10:01:54.494196 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.370259 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bsp4z"] Dec 07 10:02:01 crc kubenswrapper[4838]: E1207 10:02:01.371316 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de7aef2-84f7-4398-a557-4041f9031e31" containerName="keystone-cron" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.371331 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de7aef2-84f7-4398-a557-4041f9031e31" containerName="keystone-cron" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.371560 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de7aef2-84f7-4398-a557-4041f9031e31" containerName="keystone-cron" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.373410 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.384250 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsp4z"] Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.430197 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w668f\" (UniqueName: \"kubernetes.io/projected/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-kube-api-access-w668f\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.430284 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-catalog-content\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.430412 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-utilities\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.532224 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w668f\" (UniqueName: \"kubernetes.io/projected/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-kube-api-access-w668f\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.532299 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-catalog-content\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.532404 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-utilities\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.532790 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-catalog-content\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.532924 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-utilities\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.550642 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w668f\" (UniqueName: \"kubernetes.io/projected/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-kube-api-access-w668f\") pod \"redhat-marketplace-bsp4z\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:01 crc kubenswrapper[4838]: I1207 10:02:01.690729 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:02 crc kubenswrapper[4838]: I1207 10:02:02.284243 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsp4z"] Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.051668 4838 generic.go:334] "Generic (PLEG): container finished" podID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerID="2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863" exitCode=0 Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.051727 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsp4z" event={"ID":"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc","Type":"ContainerDied","Data":"2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863"} Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.052006 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsp4z" event={"ID":"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc","Type":"ContainerStarted","Data":"55791cd385dc492f94f3b85d50c250674c60686147698f7435b8ed5897fc5015"} Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.770478 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k2df9"] Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.774415 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.790541 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k2df9"] Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.885201 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-utilities\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.886091 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-catalog-content\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.886131 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmrmf\" (UniqueName: \"kubernetes.io/projected/6144c302-81de-4261-b78c-c74a1f8463a1-kube-api-access-xmrmf\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.987564 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-catalog-content\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.987707 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmrmf\" (UniqueName: \"kubernetes.io/projected/6144c302-81de-4261-b78c-c74a1f8463a1-kube-api-access-xmrmf\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.987826 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-utilities\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.988108 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-catalog-content\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:03 crc kubenswrapper[4838]: I1207 10:02:03.988156 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-utilities\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:04 crc kubenswrapper[4838]: I1207 10:02:04.011194 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmrmf\" (UniqueName: \"kubernetes.io/projected/6144c302-81de-4261-b78c-c74a1f8463a1-kube-api-access-xmrmf\") pod \"redhat-operators-k2df9\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:04 crc kubenswrapper[4838]: I1207 10:02:04.062383 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsp4z" event={"ID":"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc","Type":"ContainerStarted","Data":"f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a"} Dec 07 10:02:04 crc kubenswrapper[4838]: I1207 10:02:04.112862 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:04 crc kubenswrapper[4838]: I1207 10:02:04.624901 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k2df9"] Dec 07 10:02:05 crc kubenswrapper[4838]: I1207 10:02:05.071393 4838 generic.go:334] "Generic (PLEG): container finished" podID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerID="f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a" exitCode=0 Dec 07 10:02:05 crc kubenswrapper[4838]: I1207 10:02:05.071443 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsp4z" event={"ID":"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc","Type":"ContainerDied","Data":"f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a"} Dec 07 10:02:05 crc kubenswrapper[4838]: I1207 10:02:05.074900 4838 generic.go:334] "Generic (PLEG): container finished" podID="6144c302-81de-4261-b78c-c74a1f8463a1" containerID="287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834" exitCode=0 Dec 07 10:02:05 crc kubenswrapper[4838]: I1207 10:02:05.074933 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2df9" event={"ID":"6144c302-81de-4261-b78c-c74a1f8463a1","Type":"ContainerDied","Data":"287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834"} Dec 07 10:02:05 crc kubenswrapper[4838]: I1207 10:02:05.074956 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2df9" event={"ID":"6144c302-81de-4261-b78c-c74a1f8463a1","Type":"ContainerStarted","Data":"f088f58f4e25401187ce23e1c81284eac8ae27ed766bc6aa914dd200a3e99b8d"} Dec 07 10:02:06 crc kubenswrapper[4838]: I1207 10:02:06.090216 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsp4z" event={"ID":"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc","Type":"ContainerStarted","Data":"8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077"} Dec 07 10:02:06 crc kubenswrapper[4838]: I1207 10:02:06.094300 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2df9" event={"ID":"6144c302-81de-4261-b78c-c74a1f8463a1","Type":"ContainerStarted","Data":"61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e"} Dec 07 10:02:06 crc kubenswrapper[4838]: I1207 10:02:06.143441 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bsp4z" podStartSLOduration=2.743411975 podStartE2EDuration="5.143422659s" podCreationTimestamp="2025-12-07 10:02:01 +0000 UTC" firstStartedPulling="2025-12-07 10:02:03.053884291 +0000 UTC m=+3339.761203308" lastFinishedPulling="2025-12-07 10:02:05.453894965 +0000 UTC m=+3342.161213992" observedRunningTime="2025-12-07 10:02:06.118006238 +0000 UTC m=+3342.825325255" watchObservedRunningTime="2025-12-07 10:02:06.143422659 +0000 UTC m=+3342.850741676" Dec 07 10:02:09 crc kubenswrapper[4838]: I1207 10:02:09.128225 4838 generic.go:334] "Generic (PLEG): container finished" podID="6144c302-81de-4261-b78c-c74a1f8463a1" containerID="61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e" exitCode=0 Dec 07 10:02:09 crc kubenswrapper[4838]: I1207 10:02:09.128327 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2df9" event={"ID":"6144c302-81de-4261-b78c-c74a1f8463a1","Type":"ContainerDied","Data":"61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e"} Dec 07 10:02:10 crc kubenswrapper[4838]: I1207 10:02:10.138667 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2df9" event={"ID":"6144c302-81de-4261-b78c-c74a1f8463a1","Type":"ContainerStarted","Data":"5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275"} Dec 07 10:02:10 crc kubenswrapper[4838]: I1207 10:02:10.158940 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k2df9" podStartSLOduration=2.55270022 podStartE2EDuration="7.158901586s" podCreationTimestamp="2025-12-07 10:02:03 +0000 UTC" firstStartedPulling="2025-12-07 10:02:05.076702321 +0000 UTC m=+3341.784021338" lastFinishedPulling="2025-12-07 10:02:09.682903687 +0000 UTC m=+3346.390222704" observedRunningTime="2025-12-07 10:02:10.155986594 +0000 UTC m=+3346.863305611" watchObservedRunningTime="2025-12-07 10:02:10.158901586 +0000 UTC m=+3346.866220603" Dec 07 10:02:11 crc kubenswrapper[4838]: I1207 10:02:11.691445 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:11 crc kubenswrapper[4838]: I1207 10:02:11.691791 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:11 crc kubenswrapper[4838]: I1207 10:02:11.753126 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:12 crc kubenswrapper[4838]: I1207 10:02:12.163792 4838 generic.go:334] "Generic (PLEG): container finished" podID="e47e7bbd-a623-4a13-ba01-2fb7ff984b40" containerID="978dd07beeb62d54f0961e800bc1d177c4156fa25e9a2c6b5b2e196bedcd646e" exitCode=0 Dec 07 10:02:12 crc kubenswrapper[4838]: I1207 10:02:12.163855 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" event={"ID":"e47e7bbd-a623-4a13-ba01-2fb7ff984b40","Type":"ContainerDied","Data":"978dd07beeb62d54f0961e800bc1d177c4156fa25e9a2c6b5b2e196bedcd646e"} Dec 07 10:02:12 crc kubenswrapper[4838]: I1207 10:02:12.237728 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.164633 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsp4z"] Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.644241 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.770902 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-extra-config-0\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.770953 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ssh-key\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.770985 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-1\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771034 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771109 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph-nova-0\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771138 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-inventory\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771182 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-1\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771207 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-0\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771224 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-0\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771244 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-custom-ceph-combined-ca-bundle\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.771267 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hjq9\" (UniqueName: \"kubernetes.io/projected/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-kube-api-access-4hjq9\") pod \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\" (UID: \"e47e7bbd-a623-4a13-ba01-2fb7ff984b40\") " Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.799467 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-kube-api-access-4hjq9" (OuterVolumeSpecName: "kube-api-access-4hjq9") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "kube-api-access-4hjq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.801432 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.801245 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.804284 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph" (OuterVolumeSpecName: "ceph") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.812723 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.818437 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-inventory" (OuterVolumeSpecName: "inventory") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.819024 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.824730 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.826328 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.837384 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.844551 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "e47e7bbd-a623-4a13-ba01-2fb7ff984b40" (UID: "e47e7bbd-a623-4a13-ba01-2fb7ff984b40"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873104 4838 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873139 4838 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873148 4838 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873157 4838 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873169 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hjq9\" (UniqueName: \"kubernetes.io/projected/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-kube-api-access-4hjq9\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873179 4838 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873188 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873196 4838 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873208 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873216 4838 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:13 crc kubenswrapper[4838]: I1207 10:02:13.873224 4838 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e47e7bbd-a623-4a13-ba01-2fb7ff984b40-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.114091 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.114152 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.182775 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" event={"ID":"e47e7bbd-a623-4a13-ba01-2fb7ff984b40","Type":"ContainerDied","Data":"ebe54e3962626f6d7f26a1972051a2995270cc2626575a5ddc63cc5ba3181846"} Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.183123 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe54e3962626f6d7f26a1972051a2995270cc2626575a5ddc63cc5ba3181846" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.182906 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bsp4z" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="registry-server" containerID="cri-o://8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077" gracePeriod=2 Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.182804 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.660218 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.788914 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-utilities\") pod \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.789295 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-catalog-content\") pod \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.789335 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w668f\" (UniqueName: \"kubernetes.io/projected/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-kube-api-access-w668f\") pod \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\" (UID: \"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc\") " Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.790224 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-utilities" (OuterVolumeSpecName: "utilities") pod "eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" (UID: "eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.803021 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-kube-api-access-w668f" (OuterVolumeSpecName: "kube-api-access-w668f") pod "eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" (UID: "eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc"). InnerVolumeSpecName "kube-api-access-w668f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.809697 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" (UID: "eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.891863 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.891911 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:14 crc kubenswrapper[4838]: I1207 10:02:14.891925 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w668f\" (UniqueName: \"kubernetes.io/projected/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc-kube-api-access-w668f\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.166044 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k2df9" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="registry-server" probeResult="failure" output=< Dec 07 10:02:15 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 10:02:15 crc kubenswrapper[4838]: > Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.193793 4838 generic.go:334] "Generic (PLEG): container finished" podID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerID="8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077" exitCode=0 Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.193875 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsp4z" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.193882 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsp4z" event={"ID":"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc","Type":"ContainerDied","Data":"8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077"} Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.194040 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsp4z" event={"ID":"eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc","Type":"ContainerDied","Data":"55791cd385dc492f94f3b85d50c250674c60686147698f7435b8ed5897fc5015"} Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.194105 4838 scope.go:117] "RemoveContainer" containerID="8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.219459 4838 scope.go:117] "RemoveContainer" containerID="f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.236385 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsp4z"] Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.246992 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsp4z"] Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.249933 4838 scope.go:117] "RemoveContainer" containerID="2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.283506 4838 scope.go:117] "RemoveContainer" containerID="8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077" Dec 07 10:02:15 crc kubenswrapper[4838]: E1207 10:02:15.284991 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077\": container with ID starting with 8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077 not found: ID does not exist" containerID="8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.285036 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077"} err="failed to get container status \"8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077\": rpc error: code = NotFound desc = could not find container \"8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077\": container with ID starting with 8c321beb4b48c2fb03dd41842d9e6417c4d5ce9eb2920854723ac002b0cdc077 not found: ID does not exist" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.285062 4838 scope.go:117] "RemoveContainer" containerID="f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a" Dec 07 10:02:15 crc kubenswrapper[4838]: E1207 10:02:15.285624 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a\": container with ID starting with f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a not found: ID does not exist" containerID="f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.285691 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a"} err="failed to get container status \"f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a\": rpc error: code = NotFound desc = could not find container \"f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a\": container with ID starting with f99224bb36eaa1b8d4f9c8a5a22201dd1b02381e9a440613ef8fe9491fba8a6a not found: ID does not exist" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.285741 4838 scope.go:117] "RemoveContainer" containerID="2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863" Dec 07 10:02:15 crc kubenswrapper[4838]: E1207 10:02:15.286018 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863\": container with ID starting with 2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863 not found: ID does not exist" containerID="2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.286046 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863"} err="failed to get container status \"2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863\": rpc error: code = NotFound desc = could not find container \"2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863\": container with ID starting with 2966f5f90910d7ec6f87701ee5b4f4e26784ba75c17ed3db052a6f3652024863 not found: ID does not exist" Dec 07 10:02:15 crc kubenswrapper[4838]: I1207 10:02:15.626272 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" path="/var/lib/kubelet/pods/eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc/volumes" Dec 07 10:02:24 crc kubenswrapper[4838]: I1207 10:02:24.196431 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:24 crc kubenswrapper[4838]: I1207 10:02:24.297990 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:24 crc kubenswrapper[4838]: I1207 10:02:24.493479 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:02:24 crc kubenswrapper[4838]: I1207 10:02:24.493523 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:02:24 crc kubenswrapper[4838]: I1207 10:02:24.706882 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k2df9"] Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.314315 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k2df9" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="registry-server" containerID="cri-o://5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275" gracePeriod=2 Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.843941 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.932065 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmrmf\" (UniqueName: \"kubernetes.io/projected/6144c302-81de-4261-b78c-c74a1f8463a1-kube-api-access-xmrmf\") pod \"6144c302-81de-4261-b78c-c74a1f8463a1\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.932130 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-utilities\") pod \"6144c302-81de-4261-b78c-c74a1f8463a1\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.932250 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-catalog-content\") pod \"6144c302-81de-4261-b78c-c74a1f8463a1\" (UID: \"6144c302-81de-4261-b78c-c74a1f8463a1\") " Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.932794 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-utilities" (OuterVolumeSpecName: "utilities") pod "6144c302-81de-4261-b78c-c74a1f8463a1" (UID: "6144c302-81de-4261-b78c-c74a1f8463a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.938980 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6144c302-81de-4261-b78c-c74a1f8463a1-kube-api-access-xmrmf" (OuterVolumeSpecName: "kube-api-access-xmrmf") pod "6144c302-81de-4261-b78c-c74a1f8463a1" (UID: "6144c302-81de-4261-b78c-c74a1f8463a1"). InnerVolumeSpecName "kube-api-access-xmrmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.945456 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmrmf\" (UniqueName: \"kubernetes.io/projected/6144c302-81de-4261-b78c-c74a1f8463a1-kube-api-access-xmrmf\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:25 crc kubenswrapper[4838]: I1207 10:02:25.945483 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.063365 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6144c302-81de-4261-b78c-c74a1f8463a1" (UID: "6144c302-81de-4261-b78c-c74a1f8463a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.149056 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6144c302-81de-4261-b78c-c74a1f8463a1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.325766 4838 generic.go:334] "Generic (PLEG): container finished" podID="6144c302-81de-4261-b78c-c74a1f8463a1" containerID="5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275" exitCode=0 Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.325803 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2df9" event={"ID":"6144c302-81de-4261-b78c-c74a1f8463a1","Type":"ContainerDied","Data":"5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275"} Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.325879 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2df9" event={"ID":"6144c302-81de-4261-b78c-c74a1f8463a1","Type":"ContainerDied","Data":"f088f58f4e25401187ce23e1c81284eac8ae27ed766bc6aa914dd200a3e99b8d"} Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.325876 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2df9" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.325934 4838 scope.go:117] "RemoveContainer" containerID="5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.366579 4838 scope.go:117] "RemoveContainer" containerID="61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.380759 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k2df9"] Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.399902 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k2df9"] Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.422278 4838 scope.go:117] "RemoveContainer" containerID="287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.452923 4838 scope.go:117] "RemoveContainer" containerID="5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275" Dec 07 10:02:26 crc kubenswrapper[4838]: E1207 10:02:26.453931 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275\": container with ID starting with 5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275 not found: ID does not exist" containerID="5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.453968 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275"} err="failed to get container status \"5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275\": rpc error: code = NotFound desc = could not find container \"5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275\": container with ID starting with 5822c6bc6b6c24c5300c027d28239350507342224f70bdcfcea3d5e9d10bd275 not found: ID does not exist" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.453993 4838 scope.go:117] "RemoveContainer" containerID="61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e" Dec 07 10:02:26 crc kubenswrapper[4838]: E1207 10:02:26.456291 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e\": container with ID starting with 61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e not found: ID does not exist" containerID="61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.456341 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e"} err="failed to get container status \"61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e\": rpc error: code = NotFound desc = could not find container \"61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e\": container with ID starting with 61ec9a874232f20779b603aaa7e648593f93ad5c67abc164dbadd91c1552a57e not found: ID does not exist" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.456369 4838 scope.go:117] "RemoveContainer" containerID="287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834" Dec 07 10:02:26 crc kubenswrapper[4838]: E1207 10:02:26.456764 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834\": container with ID starting with 287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834 not found: ID does not exist" containerID="287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834" Dec 07 10:02:26 crc kubenswrapper[4838]: I1207 10:02:26.456801 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834"} err="failed to get container status \"287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834\": rpc error: code = NotFound desc = could not find container \"287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834\": container with ID starting with 287a69eb78f25dd99f29716abdb9cc5a97711e2d9d3eed52edf723c5d14a3834 not found: ID does not exist" Dec 07 10:02:27 crc kubenswrapper[4838]: I1207 10:02:27.626064 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" path="/var/lib/kubelet/pods/6144c302-81de-4261-b78c-c74a1f8463a1/volumes" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121074 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 07 10:02:29 crc kubenswrapper[4838]: E1207 10:02:29.121669 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="extract-utilities" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121681 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="extract-utilities" Dec 07 10:02:29 crc kubenswrapper[4838]: E1207 10:02:29.121706 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="registry-server" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121712 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="registry-server" Dec 07 10:02:29 crc kubenswrapper[4838]: E1207 10:02:29.121721 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="extract-content" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121727 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="extract-content" Dec 07 10:02:29 crc kubenswrapper[4838]: E1207 10:02:29.121740 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="extract-content" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121746 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="extract-content" Dec 07 10:02:29 crc kubenswrapper[4838]: E1207 10:02:29.121757 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e47e7bbd-a623-4a13-ba01-2fb7ff984b40" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121764 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e47e7bbd-a623-4a13-ba01-2fb7ff984b40" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 07 10:02:29 crc kubenswrapper[4838]: E1207 10:02:29.121776 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="extract-utilities" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121782 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="extract-utilities" Dec 07 10:02:29 crc kubenswrapper[4838]: E1207 10:02:29.121792 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="registry-server" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121798 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="registry-server" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121972 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="eefbfa1e-4ddf-4fee-9a9c-1a327bae03dc" containerName="registry-server" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.121992 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6144c302-81de-4261-b78c-c74a1f8463a1" containerName="registry-server" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.122002 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e47e7bbd-a623-4a13-ba01-2fb7ff984b40" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.122937 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.130471 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.130849 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.149317 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.184935 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.186379 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.192485 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.196916 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.196975 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197016 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197041 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197072 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5d85f239-60d0-45b7-b884-a3b625160cc0-ceph\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197090 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-scripts\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197131 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197147 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-dev\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197160 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197182 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197203 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197277 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfk9t\" (UniqueName: \"kubernetes.io/projected/874c3661-01ec-4c0c-8929-e05fc0fcf66e-kube-api-access-jfk9t\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197324 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197342 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197356 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-lib-modules\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197384 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197425 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197444 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197464 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197496 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-config-data\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197517 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/874c3661-01ec-4c0c-8929-e05fc0fcf66e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197534 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-run\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197568 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197584 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197600 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197623 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-sys\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197649 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-run\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197681 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197716 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197752 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmt4d\" (UniqueName: \"kubernetes.io/projected/5d85f239-60d0-45b7-b884-a3b625160cc0-kube-api-access-xmt4d\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197768 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.197802 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.275009 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300695 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300732 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-dev\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300748 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300768 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300786 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300806 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfk9t\" (UniqueName: \"kubernetes.io/projected/874c3661-01ec-4c0c-8929-e05fc0fcf66e-kube-api-access-jfk9t\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300839 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300856 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300871 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-lib-modules\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300891 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300921 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300939 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300957 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300979 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-config-data\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.300994 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/874c3661-01ec-4c0c-8929-e05fc0fcf66e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301011 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-run\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301032 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301047 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301060 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301079 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-sys\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301096 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-run\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301115 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301143 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301165 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmt4d\" (UniqueName: \"kubernetes.io/projected/5d85f239-60d0-45b7-b884-a3b625160cc0-kube-api-access-xmt4d\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301180 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301199 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301221 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301235 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301255 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301271 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301290 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5d85f239-60d0-45b7-b884-a3b625160cc0-ceph\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.301304 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-scripts\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.305757 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.305830 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.305855 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-lib-modules\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.305956 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.306111 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.306152 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.306154 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.306173 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-run\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.306290 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.307021 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.307052 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-dev\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.307122 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.307144 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.307166 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.307676 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.311532 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.312667 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.312733 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.312759 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.312841 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/874c3661-01ec-4c0c-8929-e05fc0fcf66e-run\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.312862 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5d85f239-60d0-45b7-b884-a3b625160cc0-sys\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.319216 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-scripts\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.326424 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.326555 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-config-data\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.328310 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.336420 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/874c3661-01ec-4c0c-8929-e05fc0fcf66e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.338225 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfk9t\" (UniqueName: \"kubernetes.io/projected/874c3661-01ec-4c0c-8929-e05fc0fcf66e-kube-api-access-jfk9t\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.338974 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.340497 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d85f239-60d0-45b7-b884-a3b625160cc0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.342623 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5d85f239-60d0-45b7-b884-a3b625160cc0-ceph\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.342680 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/874c3661-01ec-4c0c-8929-e05fc0fcf66e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"874c3661-01ec-4c0c-8929-e05fc0fcf66e\") " pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.351575 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmt4d\" (UniqueName: \"kubernetes.io/projected/5d85f239-60d0-45b7-b884-a3b625160cc0-kube-api-access-xmt4d\") pod \"cinder-backup-0\" (UID: \"5d85f239-60d0-45b7-b884-a3b625160cc0\") " pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.437473 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.507236 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.792390 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-gzs9g"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.793464 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.824892 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-gzs9g"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.914653 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-7e54-account-create-update-c6vjz"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.916346 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.918921 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-operator-scripts\") pod \"manila-db-create-gzs9g\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.918995 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w84qv\" (UniqueName: \"kubernetes.io/projected/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-kube-api-access-w84qv\") pod \"manila-db-create-gzs9g\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.919246 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.944189 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-7e54-account-create-update-c6vjz"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.992209 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:29 crc kubenswrapper[4838]: I1207 10:02:29.993781 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.003159 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.008458 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pw4f6" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.008630 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.009366 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.009447 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.020341 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-operator-scripts\") pod \"manila-db-create-gzs9g\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.020429 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w84qv\" (UniqueName: \"kubernetes.io/projected/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-kube-api-access-w84qv\") pod \"manila-db-create-gzs9g\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.020479 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqbbw\" (UniqueName: \"kubernetes.io/projected/92cc0e9b-6cd2-4a17-9e61-95931c154279-kube-api-access-sqbbw\") pod \"manila-7e54-account-create-update-c6vjz\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.020535 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92cc0e9b-6cd2-4a17-9e61-95931c154279-operator-scripts\") pod \"manila-7e54-account-create-update-c6vjz\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.021359 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-operator-scripts\") pod \"manila-db-create-gzs9g\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.052977 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w84qv\" (UniqueName: \"kubernetes.io/projected/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-kube-api-access-w84qv\") pod \"manila-db-create-gzs9g\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.086588 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.088097 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.094441 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.099154 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.099310 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-868f9df9df-sv4m8"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.101681 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.109443 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.113420 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.113612 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.113722 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-m24z6" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.113855 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.120248 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121420 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121459 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-config-data\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121521 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdl6b\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-kube-api-access-hdl6b\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121549 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121567 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121619 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqbbw\" (UniqueName: \"kubernetes.io/projected/92cc0e9b-6cd2-4a17-9e61-95931c154279-kube-api-access-sqbbw\") pod \"manila-7e54-account-create-update-c6vjz\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121635 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-logs\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121667 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-ceph\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121692 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-scripts\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121716 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92cc0e9b-6cd2-4a17-9e61-95931c154279-operator-scripts\") pod \"manila-7e54-account-create-update-c6vjz\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.121745 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.122739 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92cc0e9b-6cd2-4a17-9e61-95931c154279-operator-scripts\") pod \"manila-7e54-account-create-update-c6vjz\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.178482 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqbbw\" (UniqueName: \"kubernetes.io/projected/92cc0e9b-6cd2-4a17-9e61-95931c154279-kube-api-access-sqbbw\") pod \"manila-7e54-account-create-update-c6vjz\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.179439 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-868f9df9df-sv4m8"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.212981 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:30 crc kubenswrapper[4838]: E1207 10:02:30.213691 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-4zqng logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="f78ee076-a2a4-4050-8f9c-b932a81e5e08" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226368 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdl6b\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-kube-api-access-hdl6b\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226413 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-logs\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226434 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226457 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226475 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226496 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-logs\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226524 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226549 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226563 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-ceph\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226578 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-config-data\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226604 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-scripts\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226621 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-scripts\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226641 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5516b801-6088-4d82-be96-ae1efea7283d-horizon-secret-key\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226665 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226683 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226703 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226721 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226749 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zqng\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-kube-api-access-4zqng\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226774 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226793 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226873 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-config-data\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226896 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbs2d\" (UniqueName: \"kubernetes.io/projected/5516b801-6088-4d82-be96-ae1efea7283d-kube-api-access-jbs2d\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.226938 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5516b801-6088-4d82-be96-ae1efea7283d-logs\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.228156 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.229740 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-logs\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.232006 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.237185 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-scripts\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.257587 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.258014 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-ceph\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.276792 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.277435 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d55754b6f-pfkjv"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.279291 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.284723 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.285260 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdl6b\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-kube-api-access-hdl6b\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.295592 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-config-data\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.322301 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d55754b6f-pfkjv"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330073 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330135 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330173 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330192 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-config-data\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330233 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-scripts\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330260 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5516b801-6088-4d82-be96-ae1efea7283d-horizon-secret-key\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330292 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330322 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330339 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330381 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zqng\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-kube-api-access-4zqng\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330423 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330455 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbs2d\" (UniqueName: \"kubernetes.io/projected/5516b801-6088-4d82-be96-ae1efea7283d-kube-api-access-jbs2d\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330514 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5516b801-6088-4d82-be96-ae1efea7283d-logs\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330576 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-logs\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.330992 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-logs\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.332607 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5516b801-6088-4d82-be96-ae1efea7283d-logs\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.354549 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.355029 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.356063 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-config-data\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.360566 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.361723 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-scripts\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.366645 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.372415 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5516b801-6088-4d82-be96-ae1efea7283d-horizon-secret-key\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.373187 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.373830 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.396240 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.402801 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbs2d\" (UniqueName: \"kubernetes.io/projected/5516b801-6088-4d82-be96-ae1efea7283d-kube-api-access-jbs2d\") pod \"horizon-868f9df9df-sv4m8\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.412251 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.412508 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.416391 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.429502 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zqng\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-kube-api-access-4zqng\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.434877 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-scripts\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.434951 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mkj2\" (UniqueName: \"kubernetes.io/projected/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-kube-api-access-8mkj2\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.435044 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-config-data\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.435068 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-logs\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.435094 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-horizon-secret-key\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.440540 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.441327 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.449171 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.478411 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.531627 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.536528 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-scripts\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.536616 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mkj2\" (UniqueName: \"kubernetes.io/projected/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-kube-api-access-8mkj2\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.536760 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-config-data\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.536798 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-logs\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.536887 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-horizon-secret-key\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.539450 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-logs\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.540050 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-config-data\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.541278 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-scripts\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: W1207 10:02:30.550067 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d85f239_60d0_45b7_b884_a3b625160cc0.slice/crio-aa4fda1d7567148f472f93dbb5fafac93c9a5944432cffe26b8b756ac9d5fe62 WatchSource:0}: Error finding container aa4fda1d7567148f472f93dbb5fafac93c9a5944432cffe26b8b756ac9d5fe62: Status 404 returned error can't find the container with id aa4fda1d7567148f472f93dbb5fafac93c9a5944432cffe26b8b756ac9d5fe62 Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.550564 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-horizon-secret-key\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.555802 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mkj2\" (UniqueName: \"kubernetes.io/projected/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-kube-api-access-8mkj2\") pod \"horizon-7d55754b6f-pfkjv\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.602357 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.611321 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751291 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-config-data\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751372 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-ceph\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751408 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751466 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-combined-ca-bundle\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751507 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zqng\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-kube-api-access-4zqng\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751546 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-internal-tls-certs\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751618 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-logs\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.751676 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-scripts\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.752696 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-httpd-run\") pod \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\" (UID: \"f78ee076-a2a4-4050-8f9c-b932a81e5e08\") " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.752878 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.752936 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-logs" (OuterVolumeSpecName: "logs") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.753305 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-logs\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.753322 4838 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f78ee076-a2a4-4050-8f9c-b932a81e5e08-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.755618 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.757356 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-ceph" (OuterVolumeSpecName: "ceph") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.757851 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-scripts" (OuterVolumeSpecName: "scripts") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.758632 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.760481 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-kube-api-access-4zqng" (OuterVolumeSpecName: "kube-api-access-4zqng") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "kube-api-access-4zqng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.779994 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-config-data" (OuterVolumeSpecName: "config-data") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.785934 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f78ee076-a2a4-4050-8f9c-b932a81e5e08" (UID: "f78ee076-a2a4-4050-8f9c-b932a81e5e08"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.859182 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.859216 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.859249 4838 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.859262 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.859275 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zqng\" (UniqueName: \"kubernetes.io/projected/f78ee076-a2a4-4050-8f9c-b932a81e5e08-kube-api-access-4zqng\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.859286 4838 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.859294 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f78ee076-a2a4-4050-8f9c-b932a81e5e08-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:30 crc kubenswrapper[4838]: I1207 10:02:30.995939 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-gzs9g"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.030887 4838 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.089110 4838 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.124414 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-7e54-account-create-update-c6vjz"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.149702 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-868f9df9df-sv4m8"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.422330 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:31 crc kubenswrapper[4838]: W1207 10:02:31.427770 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e535673_77a0_4d31_aa8b_e64c77202c8b.slice/crio-ba3531bd72bfa0e81ab9a7b29b6bbb04575d1d188dc7e45ee83f71c9fdf47772 WatchSource:0}: Error finding container ba3531bd72bfa0e81ab9a7b29b6bbb04575d1d188dc7e45ee83f71c9fdf47772: Status 404 returned error can't find the container with id ba3531bd72bfa0e81ab9a7b29b6bbb04575d1d188dc7e45ee83f71c9fdf47772 Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.428340 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-gzs9g" event={"ID":"521c679e-583d-4ee9-827e-bb3b4a1e7e1e","Type":"ContainerStarted","Data":"2c22deb3bd2ed0df274bbf6186dd265615907151ad915644b2f0172f61681188"} Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.459104 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5d85f239-60d0-45b7-b884-a3b625160cc0","Type":"ContainerStarted","Data":"aa4fda1d7567148f472f93dbb5fafac93c9a5944432cffe26b8b756ac9d5fe62"} Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.463783 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d55754b6f-pfkjv"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.466069 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-7e54-account-create-update-c6vjz" event={"ID":"92cc0e9b-6cd2-4a17-9e61-95931c154279","Type":"ContainerStarted","Data":"4514ac173d852b948e39dec9c343c1015b4bb137bde068e0fab72a399ffbca8c"} Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.476812 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"874c3661-01ec-4c0c-8929-e05fc0fcf66e","Type":"ContainerStarted","Data":"870318dd63ad1689d7885b57773dc63dcaaabb3c12a508a3f672539e5085afc7"} Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.487360 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.499739 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868f9df9df-sv4m8" event={"ID":"5516b801-6088-4d82-be96-ae1efea7283d","Type":"ContainerStarted","Data":"98454e31d463719bc79cedcc8d9ebd6e1ae73f5516c13dbaaa8247a459d8e8eb"} Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.671580 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.671616 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.680910 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.682541 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.694291 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.694543 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.703312 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.810755 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.810833 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.810869 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.811433 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.811491 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.811546 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-logs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.811572 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.811612 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4pfs\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-kube-api-access-h4pfs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.811645 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.915831 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.915888 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.915909 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.915929 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.915970 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.916011 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-logs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.916034 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.916068 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4pfs\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-kube-api-access-h4pfs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.916095 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.917256 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-logs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.917475 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.917693 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.921707 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.924761 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.925426 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.927440 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.930617 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.941490 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4pfs\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-kube-api-access-h4pfs\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:31 crc kubenswrapper[4838]: I1207 10:02:31.968644 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.032229 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.510953 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d55754b6f-pfkjv" event={"ID":"74a02cf1-7f14-462a-8d5f-8d0df8e8828b","Type":"ContainerStarted","Data":"0103d3fe3d334ded1fb07a74392aff98d5c36a1562007f692c17875d66c82e7c"} Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.515918 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8e535673-77a0-4d31-aa8b-e64c77202c8b","Type":"ContainerStarted","Data":"ba3531bd72bfa0e81ab9a7b29b6bbb04575d1d188dc7e45ee83f71c9fdf47772"} Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.525551 4838 generic.go:334] "Generic (PLEG): container finished" podID="92cc0e9b-6cd2-4a17-9e61-95931c154279" containerID="2486309f20fa32b2866b35fdf4f049838011c55d36a58422124a648cf2fe14f8" exitCode=0 Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.525652 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-7e54-account-create-update-c6vjz" event={"ID":"92cc0e9b-6cd2-4a17-9e61-95931c154279","Type":"ContainerDied","Data":"2486309f20fa32b2866b35fdf4f049838011c55d36a58422124a648cf2fe14f8"} Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.547401 4838 generic.go:334] "Generic (PLEG): container finished" podID="521c679e-583d-4ee9-827e-bb3b4a1e7e1e" containerID="5f54bf662d2ab3b79c9131a3d2a81be12f47a4ca45e3b36a9f80497787e4f871" exitCode=0 Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.547447 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-gzs9g" event={"ID":"521c679e-583d-4ee9-827e-bb3b4a1e7e1e","Type":"ContainerDied","Data":"5f54bf662d2ab3b79c9131a3d2a81be12f47a4ca45e3b36a9f80497787e4f871"} Dec 07 10:02:32 crc kubenswrapper[4838]: I1207 10:02:32.747767 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.306939 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d55754b6f-pfkjv"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.337057 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-766695df86-47g96"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.339880 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.358062 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-766695df86-47g96"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.379521 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-scripts\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.379577 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-config-data\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.379616 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7t5j\" (UniqueName: \"kubernetes.io/projected/408b9a70-5be1-476e-a026-dbef0631c1d3-kube-api-access-c7t5j\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.379719 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-secret-key\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.379738 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/408b9a70-5be1-476e-a026-dbef0631c1d3-logs\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.379761 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-tls-certs\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.379828 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-combined-ca-bundle\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.391081 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.466100 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-868f9df9df-sv4m8"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.480808 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-secret-key\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.480850 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/408b9a70-5be1-476e-a026-dbef0631c1d3-logs\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.480873 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-tls-certs\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.480908 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-combined-ca-bundle\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.480943 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-scripts\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.480968 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-config-data\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.480995 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7t5j\" (UniqueName: \"kubernetes.io/projected/408b9a70-5be1-476e-a026-dbef0631c1d3-kube-api-access-c7t5j\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.496644 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/408b9a70-5be1-476e-a026-dbef0631c1d3-logs\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.497221 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-scripts\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.502882 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-config-data\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.513891 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-secret-key\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.518259 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-combined-ca-bundle\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.523805 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-tls-certs\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.580028 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7t5j\" (UniqueName: \"kubernetes.io/projected/408b9a70-5be1-476e-a026-dbef0631c1d3-kube-api-access-c7t5j\") pod \"horizon-766695df86-47g96\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.582126 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-76f6479d76-n4xcx"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.606438 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76f6479d76-n4xcx"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.606563 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.690754 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f78ee076-a2a4-4050-8f9c-b932a81e5e08" path="/var/lib/kubelet/pods/f78ee076-a2a4-4050-8f9c-b932a81e5e08/volumes" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.692352 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.692968 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5d85f239-60d0-45b7-b884-a3b625160cc0","Type":"ContainerStarted","Data":"2e445c8827a1bc75aae92d6d5c52aa7191bf511c8caf5392ac90e03be51ade08"} Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.692995 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5d85f239-60d0-45b7-b884-a3b625160cc0","Type":"ContainerStarted","Data":"2c2b1c2c719ba68e79479ee15a017067729e589ecac7c660b4c0548ccda4962d"} Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.693004 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8e535673-77a0-4d31-aa8b-e64c77202c8b","Type":"ContainerStarted","Data":"e19c7b650bead5f7e9b46ce6d1c644f7428f0ab74d6779fabcf47103c0c138f1"} Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.701155 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f8abbaf6-592e-422c-a132-ffb6d0a793fa","Type":"ContainerStarted","Data":"51366ffcdc4d5e844026997010ed0ab811dcf30654db9d8bbb36c0005328e0b4"} Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.715407 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gldkj\" (UniqueName: \"kubernetes.io/projected/3f6fe9f7-9f77-4da5-823b-b650957c5860-kube-api-access-gldkj\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.715652 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-horizon-tls-certs\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.715779 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-combined-ca-bundle\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.715914 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-horizon-secret-key\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.716086 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f6fe9f7-9f77-4da5-823b-b650957c5860-config-data\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.716184 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6fe9f7-9f77-4da5-823b-b650957c5860-logs\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.716398 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6fe9f7-9f77-4da5-823b-b650957c5860-scripts\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.747311 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766695df86-47g96" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.818104 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-horizon-tls-certs\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.818170 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-combined-ca-bundle\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.818202 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-horizon-secret-key\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.818241 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f6fe9f7-9f77-4da5-823b-b650957c5860-config-data\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.818265 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6fe9f7-9f77-4da5-823b-b650957c5860-logs\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.826297 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-combined-ca-bundle\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.827545 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f6fe9f7-9f77-4da5-823b-b650957c5860-config-data\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.828065 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6fe9f7-9f77-4da5-823b-b650957c5860-scripts\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.828740 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gldkj\" (UniqueName: \"kubernetes.io/projected/3f6fe9f7-9f77-4da5-823b-b650957c5860-kube-api-access-gldkj\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.829842 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6fe9f7-9f77-4da5-823b-b650957c5860-logs\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.831415 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-horizon-tls-certs\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.831836 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6fe9f7-9f77-4da5-823b-b650957c5860-scripts\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.885665 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gldkj\" (UniqueName: \"kubernetes.io/projected/3f6fe9f7-9f77-4da5-823b-b650957c5860-kube-api-access-gldkj\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.894979 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f6fe9f7-9f77-4da5-823b-b650957c5860-horizon-secret-key\") pod \"horizon-76f6479d76-n4xcx\" (UID: \"3f6fe9f7-9f77-4da5-823b-b650957c5860\") " pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:33 crc kubenswrapper[4838]: I1207 10:02:33.992053 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.507310 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.803268 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"874c3661-01ec-4c0c-8929-e05fc0fcf66e","Type":"ContainerStarted","Data":"2dd10d8d1353b300faea39ce9b29c8a6853364a0b86a94d3fc0fc022d899ada1"} Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.803662 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"874c3661-01ec-4c0c-8929-e05fc0fcf66e","Type":"ContainerStarted","Data":"51482f183b063eb931c98915b2db84d9b138d1fe7c1f0da4bd9b16768e5e073a"} Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.862546 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.884377 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=4.3418596449999995 podStartE2EDuration="5.884359836s" podCreationTimestamp="2025-12-07 10:02:29 +0000 UTC" firstStartedPulling="2025-12-07 10:02:30.561697534 +0000 UTC m=+3367.269016551" lastFinishedPulling="2025-12-07 10:02:32.104197735 +0000 UTC m=+3368.811516742" observedRunningTime="2025-12-07 10:02:34.085961846 +0000 UTC m=+3370.793280873" watchObservedRunningTime="2025-12-07 10:02:34.884359836 +0000 UTC m=+3371.591678853" Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.888349 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.394598999 podStartE2EDuration="5.888337967s" podCreationTimestamp="2025-12-07 10:02:29 +0000 UTC" firstStartedPulling="2025-12-07 10:02:30.458575818 +0000 UTC m=+3367.165894835" lastFinishedPulling="2025-12-07 10:02:32.952314786 +0000 UTC m=+3369.659633803" observedRunningTime="2025-12-07 10:02:34.880958371 +0000 UTC m=+3371.588277388" watchObservedRunningTime="2025-12-07 10:02:34.888337967 +0000 UTC m=+3371.595656984" Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.983950 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqbbw\" (UniqueName: \"kubernetes.io/projected/92cc0e9b-6cd2-4a17-9e61-95931c154279-kube-api-access-sqbbw\") pod \"92cc0e9b-6cd2-4a17-9e61-95931c154279\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.984027 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92cc0e9b-6cd2-4a17-9e61-95931c154279-operator-scripts\") pod \"92cc0e9b-6cd2-4a17-9e61-95931c154279\" (UID: \"92cc0e9b-6cd2-4a17-9e61-95931c154279\") " Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.984593 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92cc0e9b-6cd2-4a17-9e61-95931c154279-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92cc0e9b-6cd2-4a17-9e61-95931c154279" (UID: "92cc0e9b-6cd2-4a17-9e61-95931c154279"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:02:34 crc kubenswrapper[4838]: I1207 10:02:34.984735 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92cc0e9b-6cd2-4a17-9e61-95931c154279-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.019404 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92cc0e9b-6cd2-4a17-9e61-95931c154279-kube-api-access-sqbbw" (OuterVolumeSpecName: "kube-api-access-sqbbw") pod "92cc0e9b-6cd2-4a17-9e61-95931c154279" (UID: "92cc0e9b-6cd2-4a17-9e61-95931c154279"). InnerVolumeSpecName "kube-api-access-sqbbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.025622 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.086801 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqbbw\" (UniqueName: \"kubernetes.io/projected/92cc0e9b-6cd2-4a17-9e61-95931c154279-kube-api-access-sqbbw\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.167182 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76f6479d76-n4xcx"] Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.189955 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-operator-scripts\") pod \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.193122 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w84qv\" (UniqueName: \"kubernetes.io/projected/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-kube-api-access-w84qv\") pod \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\" (UID: \"521c679e-583d-4ee9-827e-bb3b4a1e7e1e\") " Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.192530 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "521c679e-583d-4ee9-827e-bb3b4a1e7e1e" (UID: "521c679e-583d-4ee9-827e-bb3b4a1e7e1e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.234064 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-kube-api-access-w84qv" (OuterVolumeSpecName: "kube-api-access-w84qv") pod "521c679e-583d-4ee9-827e-bb3b4a1e7e1e" (UID: "521c679e-583d-4ee9-827e-bb3b4a1e7e1e"). InnerVolumeSpecName "kube-api-access-w84qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.276687 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-766695df86-47g96"] Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.296055 4838 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.296092 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w84qv\" (UniqueName: \"kubernetes.io/projected/521c679e-583d-4ee9-827e-bb3b4a1e7e1e-kube-api-access-w84qv\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.918514 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-gzs9g" event={"ID":"521c679e-583d-4ee9-827e-bb3b4a1e7e1e","Type":"ContainerDied","Data":"2c22deb3bd2ed0df274bbf6186dd265615907151ad915644b2f0172f61681188"} Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.918769 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c22deb3bd2ed0df274bbf6186dd265615907151ad915644b2f0172f61681188" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.918896 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-gzs9g" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.924192 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8e535673-77a0-4d31-aa8b-e64c77202c8b","Type":"ContainerStarted","Data":"bbed4e6ff1f6556c1a4ae07904d25fa759a1f7aedf9fec55a91f55ab4d7d7de3"} Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.924367 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-log" containerID="cri-o://e19c7b650bead5f7e9b46ce6d1c644f7428f0ab74d6779fabcf47103c0c138f1" gracePeriod=30 Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.924641 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-httpd" containerID="cri-o://bbed4e6ff1f6556c1a4ae07904d25fa759a1f7aedf9fec55a91f55ab4d7d7de3" gracePeriod=30 Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.927630 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766695df86-47g96" event={"ID":"408b9a70-5be1-476e-a026-dbef0631c1d3","Type":"ContainerStarted","Data":"6b0c3f01d610dd79d28c34c005e930b3e3caa5fe96cb21bad719b5ba5b2fa5d3"} Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.931446 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-7e54-account-create-update-c6vjz" event={"ID":"92cc0e9b-6cd2-4a17-9e61-95931c154279","Type":"ContainerDied","Data":"4514ac173d852b948e39dec9c343c1015b4bb137bde068e0fab72a399ffbca8c"} Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.931475 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4514ac173d852b948e39dec9c343c1015b4bb137bde068e0fab72a399ffbca8c" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.931526 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7e54-account-create-update-c6vjz" Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.934083 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76f6479d76-n4xcx" event={"ID":"3f6fe9f7-9f77-4da5-823b-b650957c5860","Type":"ContainerStarted","Data":"fd126171cb6b32499aeee737c2c2e12b86036d5f190b34d85caa31d1f56f0c95"} Dec 07 10:02:35 crc kubenswrapper[4838]: I1207 10:02:35.939218 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f8abbaf6-592e-422c-a132-ffb6d0a793fa","Type":"ContainerStarted","Data":"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9"} Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.089983 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.08995994 podStartE2EDuration="8.08995994s" podCreationTimestamp="2025-12-07 10:02:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:02:35.987318618 +0000 UTC m=+3372.694637635" watchObservedRunningTime="2025-12-07 10:02:36.08995994 +0000 UTC m=+3372.797278957" Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.959724 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f8abbaf6-592e-422c-a132-ffb6d0a793fa","Type":"ContainerStarted","Data":"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25"} Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.960547 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-httpd" containerID="cri-o://5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25" gracePeriod=30 Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.960114 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-log" containerID="cri-o://1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9" gracePeriod=30 Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.984670 4838 generic.go:334] "Generic (PLEG): container finished" podID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerID="bbed4e6ff1f6556c1a4ae07904d25fa759a1f7aedf9fec55a91f55ab4d7d7de3" exitCode=0 Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.984696 4838 generic.go:334] "Generic (PLEG): container finished" podID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerID="e19c7b650bead5f7e9b46ce6d1c644f7428f0ab74d6779fabcf47103c0c138f1" exitCode=143 Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.984717 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8e535673-77a0-4d31-aa8b-e64c77202c8b","Type":"ContainerDied","Data":"bbed4e6ff1f6556c1a4ae07904d25fa759a1f7aedf9fec55a91f55ab4d7d7de3"} Dec 07 10:02:36 crc kubenswrapper[4838]: I1207 10:02:36.984741 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8e535673-77a0-4d31-aa8b-e64c77202c8b","Type":"ContainerDied","Data":"e19c7b650bead5f7e9b46ce6d1c644f7428f0ab74d6779fabcf47103c0c138f1"} Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.009053 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.009034436 podStartE2EDuration="6.009034436s" podCreationTimestamp="2025-12-07 10:02:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:02:37.005897368 +0000 UTC m=+3373.713216375" watchObservedRunningTime="2025-12-07 10:02:37.009034436 +0000 UTC m=+3373.716353453" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.148440 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.251433 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-config-data\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.252003 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-scripts\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.253631 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-ceph\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.253798 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdl6b\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-kube-api-access-hdl6b\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.254022 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-public-tls-certs\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.254095 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-httpd-run\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.254202 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.254638 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-combined-ca-bundle\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.254738 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-logs\") pod \"8e535673-77a0-4d31-aa8b-e64c77202c8b\" (UID: \"8e535673-77a0-4d31-aa8b-e64c77202c8b\") " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.255156 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.256011 4838 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.257360 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-logs" (OuterVolumeSpecName: "logs") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.260579 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.267582 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-ceph" (OuterVolumeSpecName: "ceph") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.278862 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-scripts" (OuterVolumeSpecName: "scripts") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.318552 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-kube-api-access-hdl6b" (OuterVolumeSpecName: "kube-api-access-hdl6b") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "kube-api-access-hdl6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.355335 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.359369 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.359393 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.359402 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdl6b\" (UniqueName: \"kubernetes.io/projected/8e535673-77a0-4d31-aa8b-e64c77202c8b-kube-api-access-hdl6b\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.359423 4838 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.359433 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.359440 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e535673-77a0-4d31-aa8b-e64c77202c8b-logs\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.378801 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-config-data" (OuterVolumeSpecName: "config-data") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.388839 4838 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.409441 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8e535673-77a0-4d31-aa8b-e64c77202c8b" (UID: "8e535673-77a0-4d31-aa8b-e64c77202c8b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.461718 4838 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.461757 4838 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.461767 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e535673-77a0-4d31-aa8b-e64c77202c8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:37 crc kubenswrapper[4838]: I1207 10:02:37.953008 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.016559 4838 generic.go:334] "Generic (PLEG): container finished" podID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerID="5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25" exitCode=0 Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.016592 4838 generic.go:334] "Generic (PLEG): container finished" podID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerID="1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9" exitCode=143 Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.016702 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.016960 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f8abbaf6-592e-422c-a132-ffb6d0a793fa","Type":"ContainerDied","Data":"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25"} Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.017011 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f8abbaf6-592e-422c-a132-ffb6d0a793fa","Type":"ContainerDied","Data":"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9"} Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.017022 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f8abbaf6-592e-422c-a132-ffb6d0a793fa","Type":"ContainerDied","Data":"51366ffcdc4d5e844026997010ed0ab811dcf30654db9d8bbb36c0005328e0b4"} Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.017038 4838 scope.go:117] "RemoveContainer" containerID="5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.047083 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8e535673-77a0-4d31-aa8b-e64c77202c8b","Type":"ContainerDied","Data":"ba3531bd72bfa0e81ab9a7b29b6bbb04575d1d188dc7e45ee83f71c9fdf47772"} Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.047177 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.092555 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-httpd-run\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.092745 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-ceph\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.093048 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-config-data\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.093103 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-logs\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.093132 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-scripts\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.093207 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-internal-tls-certs\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.093260 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4pfs\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-kube-api-access-h4pfs\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.093323 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.093364 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-combined-ca-bundle\") pod \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\" (UID: \"f8abbaf6-592e-422c-a132-ffb6d0a793fa\") " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.100906 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-logs" (OuterVolumeSpecName: "logs") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.101164 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.150164 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-ceph" (OuterVolumeSpecName: "ceph") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.154681 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-kube-api-access-h4pfs" (OuterVolumeSpecName: "kube-api-access-h4pfs") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "kube-api-access-h4pfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.217857 4838 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.217902 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.217912 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8abbaf6-592e-422c-a132-ffb6d0a793fa-logs\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.217922 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4pfs\" (UniqueName: \"kubernetes.io/projected/f8abbaf6-592e-422c-a132-ffb6d0a793fa-kube-api-access-h4pfs\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.231056 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.236192 4838 scope.go:117] "RemoveContainer" containerID="1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.245940 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.254435 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.263309 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.266012 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-scripts" (OuterVolumeSpecName: "scripts") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.292561 4838 scope.go:117] "RemoveContainer" containerID="5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.292924 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.293738 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92cc0e9b-6cd2-4a17-9e61-95931c154279" containerName="mariadb-account-create-update" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.293836 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="92cc0e9b-6cd2-4a17-9e61-95931c154279" containerName="mariadb-account-create-update" Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.293898 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-log" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.293957 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-log" Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.294024 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-httpd" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294074 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-httpd" Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.294145 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-log" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294199 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-log" Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.294252 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521c679e-583d-4ee9-827e-bb3b4a1e7e1e" containerName="mariadb-database-create" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294313 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="521c679e-583d-4ee9-827e-bb3b4a1e7e1e" containerName="mariadb-database-create" Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.294359 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25\": container with ID starting with 5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25 not found: ID does not exist" containerID="5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294412 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25"} err="failed to get container status \"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25\": rpc error: code = NotFound desc = could not find container \"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25\": container with ID starting with 5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25 not found: ID does not exist" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294440 4838 scope.go:117] "RemoveContainer" containerID="1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9" Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.294381 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-httpd" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294511 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-httpd" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294885 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-httpd" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294909 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-httpd" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294916 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="92cc0e9b-6cd2-4a17-9e61-95931c154279" containerName="mariadb-account-create-update" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294930 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" containerName="glance-log" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294943 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" containerName="glance-log" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.294952 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="521c679e-583d-4ee9-827e-bb3b4a1e7e1e" containerName="mariadb-database-create" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.295983 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: E1207 10:02:38.298485 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9\": container with ID starting with 1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9 not found: ID does not exist" containerID="1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.298519 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9"} err="failed to get container status \"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9\": rpc error: code = NotFound desc = could not find container \"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9\": container with ID starting with 1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9 not found: ID does not exist" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.298544 4838 scope.go:117] "RemoveContainer" containerID="5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.300596 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.305363 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.306552 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.317749 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25"} err="failed to get container status \"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25\": rpc error: code = NotFound desc = could not find container \"5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25\": container with ID starting with 5303739551ec8de87cc847aa655dbb7820582dcae1d6b87e24963e2a0d728c25 not found: ID does not exist" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.317793 4838 scope.go:117] "RemoveContainer" containerID="1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.318705 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9"} err="failed to get container status \"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9\": rpc error: code = NotFound desc = could not find container \"1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9\": container with ID starting with 1ee520a5d57e2aa019bf2d5b6aeb8a51a8ca30db615bb506d58e017649b4b2f9 not found: ID does not exist" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.318743 4838 scope.go:117] "RemoveContainer" containerID="bbed4e6ff1f6556c1a4ae07904d25fa759a1f7aedf9fec55a91f55ab4d7d7de3" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.333901 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.344759 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.344793 4838 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.344889 4838 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.344903 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.401548 4838 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.403156 4838 scope.go:117] "RemoveContainer" containerID="e19c7b650bead5f7e9b46ce6d1c644f7428f0ab74d6779fabcf47103c0c138f1" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.433945 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-config-data" (OuterVolumeSpecName: "config-data") pod "f8abbaf6-592e-422c-a132-ffb6d0a793fa" (UID: "f8abbaf6-592e-422c-a132-ffb6d0a793fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446114 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-config-data\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446173 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8f155e0-a710-40b0-af6b-157bb7e25854-ceph\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446191 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8f155e0-a710-40b0-af6b-157bb7e25854-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446209 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446231 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446250 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-scripts\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446286 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f155e0-a710-40b0-af6b-157bb7e25854-logs\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446302 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446369 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zhl9\" (UniqueName: \"kubernetes.io/projected/f8f155e0-a710-40b0-af6b-157bb7e25854-kube-api-access-7zhl9\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446414 4838 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.446426 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8abbaf6-592e-422c-a132-ffb6d0a793fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549269 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f155e0-a710-40b0-af6b-157bb7e25854-logs\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549309 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549391 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zhl9\" (UniqueName: \"kubernetes.io/projected/f8f155e0-a710-40b0-af6b-157bb7e25854-kube-api-access-7zhl9\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549455 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-config-data\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549483 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8f155e0-a710-40b0-af6b-157bb7e25854-ceph\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549498 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8f155e0-a710-40b0-af6b-157bb7e25854-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549513 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549531 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.549552 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-scripts\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.552334 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.552540 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8f155e0-a710-40b0-af6b-157bb7e25854-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.552586 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f155e0-a710-40b0-af6b-157bb7e25854-logs\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.558774 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-scripts\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.583070 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-config-data\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.587909 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f8f155e0-a710-40b0-af6b-157bb7e25854-ceph\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.601781 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.602103 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f155e0-a710-40b0-af6b-157bb7e25854-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.622642 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zhl9\" (UniqueName: \"kubernetes.io/projected/f8f155e0-a710-40b0-af6b-157bb7e25854-kube-api-access-7zhl9\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.701875 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f8f155e0-a710-40b0-af6b-157bb7e25854\") " pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.712173 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.727790 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.778791 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.780338 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.785301 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.785514 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.804418 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.928051 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.969909 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970325 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970369 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970396 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57a396f6-7966-45b4-a4a5-4f91e688eaaa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970452 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970508 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxgc2\" (UniqueName: \"kubernetes.io/projected/57a396f6-7966-45b4-a4a5-4f91e688eaaa-kube-api-access-hxgc2\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970548 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a396f6-7966-45b4-a4a5-4f91e688eaaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970656 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:38 crc kubenswrapper[4838]: I1207 10:02:38.970710 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57a396f6-7966-45b4-a4a5-4f91e688eaaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072379 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072461 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxgc2\" (UniqueName: \"kubernetes.io/projected/57a396f6-7966-45b4-a4a5-4f91e688eaaa-kube-api-access-hxgc2\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072487 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a396f6-7966-45b4-a4a5-4f91e688eaaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072577 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072632 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57a396f6-7966-45b4-a4a5-4f91e688eaaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072655 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072704 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072729 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.072763 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57a396f6-7966-45b4-a4a5-4f91e688eaaa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.074662 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.075079 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a396f6-7966-45b4-a4a5-4f91e688eaaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.075333 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57a396f6-7966-45b4-a4a5-4f91e688eaaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.082546 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57a396f6-7966-45b4-a4a5-4f91e688eaaa-ceph\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.084308 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.087101 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.088526 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.096004 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxgc2\" (UniqueName: \"kubernetes.io/projected/57a396f6-7966-45b4-a4a5-4f91e688eaaa-kube-api-access-hxgc2\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.096467 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a396f6-7966-45b4-a4a5-4f91e688eaaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.103784 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"57a396f6-7966-45b4-a4a5-4f91e688eaaa\") " pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.408107 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.439486 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.643609 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e535673-77a0-4d31-aa8b-e64c77202c8b" path="/var/lib/kubelet/pods/8e535673-77a0-4d31-aa8b-e64c77202c8b/volumes" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.644609 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8abbaf6-592e-422c-a132-ffb6d0a793fa" path="/var/lib/kubelet/pods/f8abbaf6-592e-422c-a132-ffb6d0a793fa/volumes" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.755013 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.763415 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 07 10:02:39 crc kubenswrapper[4838]: I1207 10:02:39.805125 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-volume-volume1-0" podUID="874c3661-01ec-4c0c-8929-e05fc0fcf66e" containerName="cinder-volume" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 10:02:39 crc kubenswrapper[4838]: W1207 10:02:39.813326 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8f155e0_a710_40b0_af6b_157bb7e25854.slice/crio-d6422025849aaaccd4f517be63799b7c96ce0c580e78ec965c368faafd57b40c WatchSource:0}: Error finding container d6422025849aaaccd4f517be63799b7c96ce0c580e78ec965c368faafd57b40c: Status 404 returned error can't find the container with id d6422025849aaaccd4f517be63799b7c96ce0c580e78ec965c368faafd57b40c Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.104461 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8f155e0-a710-40b0-af6b-157bb7e25854","Type":"ContainerStarted","Data":"d6422025849aaaccd4f517be63799b7c96ce0c580e78ec965c368faafd57b40c"} Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.149445 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.439693 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-gfljj"] Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.441215 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.447930 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-gfljj"] Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.478268 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-j6gcb" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.478667 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.526738 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-job-config-data\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.526791 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-combined-ca-bundle\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.526857 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-config-data\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.526916 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwhrh\" (UniqueName: \"kubernetes.io/projected/fbd56671-dd88-4a32-9337-af3ee0c5c355-kube-api-access-hwhrh\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.628293 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-job-config-data\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.628341 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-combined-ca-bundle\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.628362 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-config-data\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.628427 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwhrh\" (UniqueName: \"kubernetes.io/projected/fbd56671-dd88-4a32-9337-af3ee0c5c355-kube-api-access-hwhrh\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.635220 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-combined-ca-bundle\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.658008 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwhrh\" (UniqueName: \"kubernetes.io/projected/fbd56671-dd88-4a32-9337-af3ee0c5c355-kube-api-access-hwhrh\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.668651 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-config-data\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.682450 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-job-config-data\") pod \"manila-db-sync-gfljj\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:40 crc kubenswrapper[4838]: I1207 10:02:40.840314 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gfljj" Dec 07 10:02:41 crc kubenswrapper[4838]: I1207 10:02:41.140049 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8f155e0-a710-40b0-af6b-157bb7e25854","Type":"ContainerStarted","Data":"34f80a361ed4788c5abc5ee168dd2c978006e8532891650e3a07b6170dc13a1f"} Dec 07 10:02:41 crc kubenswrapper[4838]: I1207 10:02:41.154149 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"57a396f6-7966-45b4-a4a5-4f91e688eaaa","Type":"ContainerStarted","Data":"40c1e7ce00b79842b8338a8870e453b6a8a2582c152ac77df5d9f629992cae17"} Dec 07 10:02:41 crc kubenswrapper[4838]: I1207 10:02:41.673548 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-gfljj"] Dec 07 10:02:41 crc kubenswrapper[4838]: W1207 10:02:41.693106 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbd56671_dd88_4a32_9337_af3ee0c5c355.slice/crio-3ff214b29d5a2278e6c8ca013264996f0fb1091b276a7527899a59acc0a51d8a WatchSource:0}: Error finding container 3ff214b29d5a2278e6c8ca013264996f0fb1091b276a7527899a59acc0a51d8a: Status 404 returned error can't find the container with id 3ff214b29d5a2278e6c8ca013264996f0fb1091b276a7527899a59acc0a51d8a Dec 07 10:02:42 crc kubenswrapper[4838]: I1207 10:02:42.169590 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"57a396f6-7966-45b4-a4a5-4f91e688eaaa","Type":"ContainerStarted","Data":"301a994b8cb68d9a663628aa5caaabfba8dbf8d5f6582c25c2a9093140e1bf6e"} Dec 07 10:02:42 crc kubenswrapper[4838]: I1207 10:02:42.173096 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gfljj" event={"ID":"fbd56671-dd88-4a32-9337-af3ee0c5c355","Type":"ContainerStarted","Data":"3ff214b29d5a2278e6c8ca013264996f0fb1091b276a7527899a59acc0a51d8a"} Dec 07 10:02:42 crc kubenswrapper[4838]: I1207 10:02:42.175098 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8f155e0-a710-40b0-af6b-157bb7e25854","Type":"ContainerStarted","Data":"8434de5a74ba94ac8a96ab93413ac60b4ee1e031ae7f994d94124569be0a3e02"} Dec 07 10:02:42 crc kubenswrapper[4838]: I1207 10:02:42.202657 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.202639267 podStartE2EDuration="4.202639267s" podCreationTimestamp="2025-12-07 10:02:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:02:42.196160016 +0000 UTC m=+3378.903479033" watchObservedRunningTime="2025-12-07 10:02:42.202639267 +0000 UTC m=+3378.909958284" Dec 07 10:02:43 crc kubenswrapper[4838]: I1207 10:02:43.196467 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"57a396f6-7966-45b4-a4a5-4f91e688eaaa","Type":"ContainerStarted","Data":"b202211e2b3e224df01ad254232f91a89fbaf5984ec518ec35079a446fc26d3b"} Dec 07 10:02:43 crc kubenswrapper[4838]: I1207 10:02:43.671790 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.671766846 podStartE2EDuration="5.671766846s" podCreationTimestamp="2025-12-07 10:02:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:02:43.231552378 +0000 UTC m=+3379.938871395" watchObservedRunningTime="2025-12-07 10:02:43.671766846 +0000 UTC m=+3380.379085863" Dec 07 10:02:44 crc kubenswrapper[4838]: I1207 10:02:44.444243 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 07 10:02:48 crc kubenswrapper[4838]: I1207 10:02:48.928296 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 07 10:02:48 crc kubenswrapper[4838]: I1207 10:02:48.928805 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 07 10:02:48 crc kubenswrapper[4838]: I1207 10:02:48.969024 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 07 10:02:48 crc kubenswrapper[4838]: I1207 10:02:48.980765 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 07 10:02:49 crc kubenswrapper[4838]: I1207 10:02:49.251617 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 07 10:02:49 crc kubenswrapper[4838]: I1207 10:02:49.251657 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 07 10:02:49 crc kubenswrapper[4838]: I1207 10:02:49.410444 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:49 crc kubenswrapper[4838]: I1207 10:02:49.410495 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:49 crc kubenswrapper[4838]: I1207 10:02:49.455109 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:49 crc kubenswrapper[4838]: I1207 10:02:49.477693 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:50 crc kubenswrapper[4838]: I1207 10:02:50.259879 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:50 crc kubenswrapper[4838]: I1207 10:02:50.259926 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:52 crc kubenswrapper[4838]: I1207 10:02:52.850050 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 07 10:02:52 crc kubenswrapper[4838]: I1207 10:02:52.850455 4838 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 10:02:52 crc kubenswrapper[4838]: I1207 10:02:52.867373 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 07 10:02:52 crc kubenswrapper[4838]: I1207 10:02:52.996523 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:52 crc kubenswrapper[4838]: I1207 10:02:52.996620 4838 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 10:02:52 crc kubenswrapper[4838]: I1207 10:02:52.997375 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 07 10:02:54 crc kubenswrapper[4838]: E1207 10:02:54.491353 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-manila-api@sha256:68714e821f8e4e2d905d6e5bc7fb2e713a24c02db48901fb2a11d57b80f6c584" Dec 07 10:02:54 crc kubenswrapper[4838]: E1207 10:02:54.492063 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manila-db-sync,Image:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:68714e821f8e4e2d905d6e5bc7fb2e713a24c02db48901fb2a11d57b80f6c584,Command:[/bin/bash],Args:[-c sleep 0 && /usr/bin/manila-manage --config-dir /etc/manila/manila.conf.d db sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:job-config-data,ReadOnly:true,MountPath:/etc/manila/manila.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwhrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42429,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42429,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-db-sync-gfljj_openstack(fbd56671-dd88-4a32-9337-af3ee0c5c355): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 10:02:54 crc kubenswrapper[4838]: I1207 10:02:54.492807 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:02:54 crc kubenswrapper[4838]: I1207 10:02:54.492896 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:02:54 crc kubenswrapper[4838]: I1207 10:02:54.492946 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 10:02:54 crc kubenswrapper[4838]: E1207 10:02:54.493235 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manila-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/manila-db-sync-gfljj" podUID="fbd56671-dd88-4a32-9337-af3ee0c5c355" Dec 07 10:02:54 crc kubenswrapper[4838]: I1207 10:02:54.493717 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f7e06fe287c32a6d89d0c626a75199a006e5e2536c384a448e44b3eb61037c0"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 10:02:54 crc kubenswrapper[4838]: I1207 10:02:54.493773 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://9f7e06fe287c32a6d89d0c626a75199a006e5e2536c384a448e44b3eb61037c0" gracePeriod=600 Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.327553 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766695df86-47g96" event={"ID":"408b9a70-5be1-476e-a026-dbef0631c1d3","Type":"ContainerStarted","Data":"7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0"} Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.333947 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="9f7e06fe287c32a6d89d0c626a75199a006e5e2536c384a448e44b3eb61037c0" exitCode=0 Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.334006 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"9f7e06fe287c32a6d89d0c626a75199a006e5e2536c384a448e44b3eb61037c0"} Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.334029 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3"} Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.334044 4838 scope.go:117] "RemoveContainer" containerID="1bf62f51de3a93ddc477b885a74c84fccea261dbeedd8de66007fd67a52af80f" Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.345182 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868f9df9df-sv4m8" event={"ID":"5516b801-6088-4d82-be96-ae1efea7283d","Type":"ContainerStarted","Data":"3b2d39e36a4c7493e49f3d690d6dbf8645f0f216d0f81fd814391b65c2bec238"} Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.347037 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d55754b6f-pfkjv" event={"ID":"74a02cf1-7f14-462a-8d5f-8d0df8e8828b","Type":"ContainerStarted","Data":"6ce0ffbb8cece0655b084ff3ef7b743ad39d153ac9c690af4c46750999125d11"} Dec 07 10:02:55 crc kubenswrapper[4838]: I1207 10:02:55.349029 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76f6479d76-n4xcx" event={"ID":"3f6fe9f7-9f77-4da5-823b-b650957c5860","Type":"ContainerStarted","Data":"c7ad16690b4a59f971b1c404d196e1e2600332a91d29851934cc667ba0ea9334"} Dec 07 10:02:55 crc kubenswrapper[4838]: E1207 10:02:55.349982 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manila-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-manila-api@sha256:68714e821f8e4e2d905d6e5bc7fb2e713a24c02db48901fb2a11d57b80f6c584\\\"\"" pod="openstack/manila-db-sync-gfljj" podUID="fbd56671-dd88-4a32-9337-af3ee0c5c355" Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.357634 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76f6479d76-n4xcx" event={"ID":"3f6fe9f7-9f77-4da5-823b-b650957c5860","Type":"ContainerStarted","Data":"fc3c0244b0d643a6d0692b306096576725b80e9ab3b9154cbb6dcb43d0441017"} Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.360363 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d55754b6f-pfkjv" event={"ID":"74a02cf1-7f14-462a-8d5f-8d0df8e8828b","Type":"ContainerStarted","Data":"38204a3d8e26525ee3d411dd064eeabcc1e8145ef6f662d2fc3fc5f311f00bf2"} Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.360491 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d55754b6f-pfkjv" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon-log" containerID="cri-o://6ce0ffbb8cece0655b084ff3ef7b743ad39d153ac9c690af4c46750999125d11" gracePeriod=30 Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.360701 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d55754b6f-pfkjv" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon" containerID="cri-o://38204a3d8e26525ee3d411dd064eeabcc1e8145ef6f662d2fc3fc5f311f00bf2" gracePeriod=30 Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.368470 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766695df86-47g96" event={"ID":"408b9a70-5be1-476e-a026-dbef0631c1d3","Type":"ContainerStarted","Data":"5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5"} Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.386080 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868f9df9df-sv4m8" event={"ID":"5516b801-6088-4d82-be96-ae1efea7283d","Type":"ContainerStarted","Data":"e10c53d56470d945c083cadd2dc46e67f31a74b422934c01ebb00c98b6c97a1f"} Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.386212 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-868f9df9df-sv4m8" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon-log" containerID="cri-o://3b2d39e36a4c7493e49f3d690d6dbf8645f0f216d0f81fd814391b65c2bec238" gracePeriod=30 Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.386292 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-868f9df9df-sv4m8" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon" containerID="cri-o://e10c53d56470d945c083cadd2dc46e67f31a74b422934c01ebb00c98b6c97a1f" gracePeriod=30 Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.399599 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-76f6479d76-n4xcx" podStartSLOduration=4.056740881 podStartE2EDuration="23.399528159s" podCreationTimestamp="2025-12-07 10:02:33 +0000 UTC" firstStartedPulling="2025-12-07 10:02:35.255644135 +0000 UTC m=+3371.962963142" lastFinishedPulling="2025-12-07 10:02:54.598431403 +0000 UTC m=+3391.305750420" observedRunningTime="2025-12-07 10:02:56.391370811 +0000 UTC m=+3393.098689828" watchObservedRunningTime="2025-12-07 10:02:56.399528159 +0000 UTC m=+3393.106847176" Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.423728 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d55754b6f-pfkjv" podStartSLOduration=3.300316123 podStartE2EDuration="26.423711476s" podCreationTimestamp="2025-12-07 10:02:30 +0000 UTC" firstStartedPulling="2025-12-07 10:02:31.500947925 +0000 UTC m=+3368.208266942" lastFinishedPulling="2025-12-07 10:02:54.624343278 +0000 UTC m=+3391.331662295" observedRunningTime="2025-12-07 10:02:56.419615821 +0000 UTC m=+3393.126934838" watchObservedRunningTime="2025-12-07 10:02:56.423711476 +0000 UTC m=+3393.131030493" Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.477764 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-868f9df9df-sv4m8" podStartSLOduration=3.2201041679999998 podStartE2EDuration="26.477742148s" podCreationTimestamp="2025-12-07 10:02:30 +0000 UTC" firstStartedPulling="2025-12-07 10:02:31.339848937 +0000 UTC m=+3368.047167954" lastFinishedPulling="2025-12-07 10:02:54.597486917 +0000 UTC m=+3391.304805934" observedRunningTime="2025-12-07 10:02:56.457024238 +0000 UTC m=+3393.164343255" watchObservedRunningTime="2025-12-07 10:02:56.477742148 +0000 UTC m=+3393.185061165" Dec 07 10:02:56 crc kubenswrapper[4838]: I1207 10:02:56.502336 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-766695df86-47g96" podStartSLOduration=4.171848322 podStartE2EDuration="23.502318715s" podCreationTimestamp="2025-12-07 10:02:33 +0000 UTC" firstStartedPulling="2025-12-07 10:02:35.270244494 +0000 UTC m=+3371.977563511" lastFinishedPulling="2025-12-07 10:02:54.600714887 +0000 UTC m=+3391.308033904" observedRunningTime="2025-12-07 10:02:56.497361747 +0000 UTC m=+3393.204680764" watchObservedRunningTime="2025-12-07 10:02:56.502318715 +0000 UTC m=+3393.209637722" Dec 07 10:03:00 crc kubenswrapper[4838]: I1207 10:03:00.450263 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:03:00 crc kubenswrapper[4838]: I1207 10:03:00.603077 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.526826 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rqfvz"] Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.529421 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.539290 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqfvz"] Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.630368 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-utilities\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.630475 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-catalog-content\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.630567 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfnqj\" (UniqueName: \"kubernetes.io/projected/d6ca550b-c045-4f06-98de-197e82bc35b1-kube-api-access-gfnqj\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.731783 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-utilities\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.731865 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-catalog-content\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.731923 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfnqj\" (UniqueName: \"kubernetes.io/projected/d6ca550b-c045-4f06-98de-197e82bc35b1-kube-api-access-gfnqj\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.732639 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-catalog-content\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.733007 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-utilities\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.755253 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfnqj\" (UniqueName: \"kubernetes.io/projected/d6ca550b-c045-4f06-98de-197e82bc35b1-kube-api-access-gfnqj\") pod \"certified-operators-rqfvz\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:01 crc kubenswrapper[4838]: I1207 10:03:01.862221 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:02 crc kubenswrapper[4838]: I1207 10:03:02.499922 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqfvz"] Dec 07 10:03:03 crc kubenswrapper[4838]: I1207 10:03:03.437795 4838 generic.go:334] "Generic (PLEG): container finished" podID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerID="01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30" exitCode=0 Dec 07 10:03:03 crc kubenswrapper[4838]: I1207 10:03:03.437854 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqfvz" event={"ID":"d6ca550b-c045-4f06-98de-197e82bc35b1","Type":"ContainerDied","Data":"01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30"} Dec 07 10:03:03 crc kubenswrapper[4838]: I1207 10:03:03.438032 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqfvz" event={"ID":"d6ca550b-c045-4f06-98de-197e82bc35b1","Type":"ContainerStarted","Data":"6252f72db987d42f4a17b896aff959a777f8b54a3d7689e301e13bfe7c6fcc9e"} Dec 07 10:03:03 crc kubenswrapper[4838]: I1207 10:03:03.748796 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-766695df86-47g96" Dec 07 10:03:03 crc kubenswrapper[4838]: I1207 10:03:03.749284 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-766695df86-47g96" Dec 07 10:03:03 crc kubenswrapper[4838]: I1207 10:03:03.993232 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:03:03 crc kubenswrapper[4838]: I1207 10:03:03.993534 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:03:04 crc kubenswrapper[4838]: I1207 10:03:04.449077 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqfvz" event={"ID":"d6ca550b-c045-4f06-98de-197e82bc35b1","Type":"ContainerStarted","Data":"6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77"} Dec 07 10:03:05 crc kubenswrapper[4838]: I1207 10:03:05.458177 4838 generic.go:334] "Generic (PLEG): container finished" podID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerID="6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77" exitCode=0 Dec 07 10:03:05 crc kubenswrapper[4838]: I1207 10:03:05.458483 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqfvz" event={"ID":"d6ca550b-c045-4f06-98de-197e82bc35b1","Type":"ContainerDied","Data":"6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77"} Dec 07 10:03:06 crc kubenswrapper[4838]: I1207 10:03:06.486390 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqfvz" event={"ID":"d6ca550b-c045-4f06-98de-197e82bc35b1","Type":"ContainerStarted","Data":"f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530"} Dec 07 10:03:06 crc kubenswrapper[4838]: I1207 10:03:06.517252 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rqfvz" podStartSLOduration=3.098501304 podStartE2EDuration="5.517234422s" podCreationTimestamp="2025-12-07 10:03:01 +0000 UTC" firstStartedPulling="2025-12-07 10:03:03.439549275 +0000 UTC m=+3400.146868292" lastFinishedPulling="2025-12-07 10:03:05.858282393 +0000 UTC m=+3402.565601410" observedRunningTime="2025-12-07 10:03:06.514805654 +0000 UTC m=+3403.222124681" watchObservedRunningTime="2025-12-07 10:03:06.517234422 +0000 UTC m=+3403.224553439" Dec 07 10:03:08 crc kubenswrapper[4838]: I1207 10:03:08.511515 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gfljj" event={"ID":"fbd56671-dd88-4a32-9337-af3ee0c5c355","Type":"ContainerStarted","Data":"61de118705e9018835413e44adb711b583f672aff9b1abb7f49eec9a8df63604"} Dec 07 10:03:08 crc kubenswrapper[4838]: I1207 10:03:08.560179 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-gfljj" podStartSLOduration=3.091683284 podStartE2EDuration="28.560159745s" podCreationTimestamp="2025-12-07 10:02:40 +0000 UTC" firstStartedPulling="2025-12-07 10:02:41.695442635 +0000 UTC m=+3378.402761652" lastFinishedPulling="2025-12-07 10:03:07.163919086 +0000 UTC m=+3403.871238113" observedRunningTime="2025-12-07 10:03:08.554162927 +0000 UTC m=+3405.261481944" watchObservedRunningTime="2025-12-07 10:03:08.560159745 +0000 UTC m=+3405.267478762" Dec 07 10:03:11 crc kubenswrapper[4838]: I1207 10:03:11.862951 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:11 crc kubenswrapper[4838]: I1207 10:03:11.864306 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:12 crc kubenswrapper[4838]: I1207 10:03:12.919431 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rqfvz" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="registry-server" probeResult="failure" output=< Dec 07 10:03:12 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 10:03:12 crc kubenswrapper[4838]: > Dec 07 10:03:13 crc kubenswrapper[4838]: I1207 10:03:13.757971 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Dec 07 10:03:13 crc kubenswrapper[4838]: I1207 10:03:13.995459 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-76f6479d76-n4xcx" podUID="3f6fe9f7-9f77-4da5-823b-b650957c5860" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.241:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.241:8443: connect: connection refused" Dec 07 10:03:20 crc kubenswrapper[4838]: I1207 10:03:20.627545 4838 generic.go:334] "Generic (PLEG): container finished" podID="fbd56671-dd88-4a32-9337-af3ee0c5c355" containerID="61de118705e9018835413e44adb711b583f672aff9b1abb7f49eec9a8df63604" exitCode=0 Dec 07 10:03:20 crc kubenswrapper[4838]: I1207 10:03:20.627999 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gfljj" event={"ID":"fbd56671-dd88-4a32-9337-af3ee0c5c355","Type":"ContainerDied","Data":"61de118705e9018835413e44adb711b583f672aff9b1abb7f49eec9a8df63604"} Dec 07 10:03:21 crc kubenswrapper[4838]: I1207 10:03:21.922711 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:21 crc kubenswrapper[4838]: I1207 10:03:21.996072 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.131236 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gfljj" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.169588 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rqfvz"] Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.257210 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-job-config-data\") pod \"fbd56671-dd88-4a32-9337-af3ee0c5c355\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.257320 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-combined-ca-bundle\") pod \"fbd56671-dd88-4a32-9337-af3ee0c5c355\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.257355 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwhrh\" (UniqueName: \"kubernetes.io/projected/fbd56671-dd88-4a32-9337-af3ee0c5c355-kube-api-access-hwhrh\") pod \"fbd56671-dd88-4a32-9337-af3ee0c5c355\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.257416 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-config-data\") pod \"fbd56671-dd88-4a32-9337-af3ee0c5c355\" (UID: \"fbd56671-dd88-4a32-9337-af3ee0c5c355\") " Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.267656 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "fbd56671-dd88-4a32-9337-af3ee0c5c355" (UID: "fbd56671-dd88-4a32-9337-af3ee0c5c355"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.269573 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-config-data" (OuterVolumeSpecName: "config-data") pod "fbd56671-dd88-4a32-9337-af3ee0c5c355" (UID: "fbd56671-dd88-4a32-9337-af3ee0c5c355"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.270417 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd56671-dd88-4a32-9337-af3ee0c5c355-kube-api-access-hwhrh" (OuterVolumeSpecName: "kube-api-access-hwhrh") pod "fbd56671-dd88-4a32-9337-af3ee0c5c355" (UID: "fbd56671-dd88-4a32-9337-af3ee0c5c355"). InnerVolumeSpecName "kube-api-access-hwhrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.299622 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbd56671-dd88-4a32-9337-af3ee0c5c355" (UID: "fbd56671-dd88-4a32-9337-af3ee0c5c355"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.359336 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.359365 4838 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.359375 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd56671-dd88-4a32-9337-af3ee0c5c355-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.359383 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwhrh\" (UniqueName: \"kubernetes.io/projected/fbd56671-dd88-4a32-9337-af3ee0c5c355-kube-api-access-hwhrh\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.675352 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-gfljj" Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.675329 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-gfljj" event={"ID":"fbd56671-dd88-4a32-9337-af3ee0c5c355","Type":"ContainerDied","Data":"3ff214b29d5a2278e6c8ca013264996f0fb1091b276a7527899a59acc0a51d8a"} Dec 07 10:03:22 crc kubenswrapper[4838]: I1207 10:03:22.675447 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ff214b29d5a2278e6c8ca013264996f0fb1091b276a7527899a59acc0a51d8a" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.011696 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:23 crc kubenswrapper[4838]: E1207 10:03:23.012155 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd56671-dd88-4a32-9337-af3ee0c5c355" containerName="manila-db-sync" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.012167 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd56671-dd88-4a32-9337-af3ee0c5c355" containerName="manila-db-sync" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.012377 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbd56671-dd88-4a32-9337-af3ee0c5c355" containerName="manila-db-sync" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.013333 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.020664 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-j6gcb" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.022472 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.038295 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.038323 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.038440 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.043981 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.048670 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.049692 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.077990 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078069 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4gwg\" (UniqueName: \"kubernetes.io/projected/e0929baa-19bc-4809-95b8-22daa12417f1-kube-api-access-x4gwg\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078102 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078125 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0929baa-19bc-4809-95b8-22daa12417f1-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078164 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078186 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078223 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-scripts\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078276 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078381 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078411 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-ceph\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078444 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078484 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc7zw\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-kube-api-access-bc7zw\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078527 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.078553 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-scripts\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.082893 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.179930 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-scripts\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180163 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180263 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180391 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4gwg\" (UniqueName: \"kubernetes.io/projected/e0929baa-19bc-4809-95b8-22daa12417f1-kube-api-access-x4gwg\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180527 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180643 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0929baa-19bc-4809-95b8-22daa12417f1-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180759 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180893 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0929baa-19bc-4809-95b8-22daa12417f1-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.180933 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.181079 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.181209 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-scripts\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.182210 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.182355 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.182383 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-ceph\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.182408 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.182444 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc7zw\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-kube-api-access-bc7zw\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.182483 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.191501 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-ceph\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.192048 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.193049 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.193699 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-scripts\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.194272 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-scripts\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.196878 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.202352 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.202937 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.203422 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.233527 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4gwg\" (UniqueName: \"kubernetes.io/projected/e0929baa-19bc-4809-95b8-22daa12417f1-kube-api-access-x4gwg\") pod \"manila-scheduler-0\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.244602 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc7zw\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-kube-api-access-bc7zw\") pod \"manila-share-share1-0\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.333306 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.363687 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.465996 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7944d6f96c-r7hkg"] Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.468176 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.489847 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7944d6f96c-r7hkg"] Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.604887 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-openstack-edpm-ipam\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.604962 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq6hw\" (UniqueName: \"kubernetes.io/projected/2e347003-e564-4cf3-b006-9b4b3daf4044-kube-api-access-wq6hw\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.604990 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-ovsdbserver-sb\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.605056 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-config\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.605090 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-ovsdbserver-nb\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.605142 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-dns-svc\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.720619 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rqfvz" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="registry-server" containerID="cri-o://f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530" gracePeriod=2 Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.726502 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.734624 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-config\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.734703 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-ovsdbserver-nb\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.734855 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-dns-svc\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.734901 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-openstack-edpm-ipam\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.735002 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq6hw\" (UniqueName: \"kubernetes.io/projected/2e347003-e564-4cf3-b006-9b4b3daf4044-kube-api-access-wq6hw\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.735033 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-ovsdbserver-sb\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.736738 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-ovsdbserver-sb\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.736742 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-config\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.737259 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-ovsdbserver-nb\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.738456 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-openstack-edpm-ipam\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.739174 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.750712 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e347003-e564-4cf3-b006-9b4b3daf4044-dns-svc\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.753415 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.782398 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.813862 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq6hw\" (UniqueName: \"kubernetes.io/projected/2e347003-e564-4cf3-b006-9b4b3daf4044-kube-api-access-wq6hw\") pod \"dnsmasq-dns-7944d6f96c-r7hkg\" (UID: \"2e347003-e564-4cf3-b006-9b4b3daf4044\") " pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.836869 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-scripts\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.836919 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d78ce4-380d-40d0-8f5f-c261776c397c-etc-machine-id\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.836955 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1d78ce4-380d-40d0-8f5f-c261776c397c-logs\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.836999 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.837031 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdpxv\" (UniqueName: \"kubernetes.io/projected/c1d78ce4-380d-40d0-8f5f-c261776c397c-kube-api-access-vdpxv\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.837090 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.837106 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data-custom\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.906448 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.938405 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d78ce4-380d-40d0-8f5f-c261776c397c-etc-machine-id\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.938718 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1d78ce4-380d-40d0-8f5f-c261776c397c-logs\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.938767 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.938791 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdpxv\" (UniqueName: \"kubernetes.io/projected/c1d78ce4-380d-40d0-8f5f-c261776c397c-kube-api-access-vdpxv\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.938871 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.938894 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data-custom\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.938952 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-scripts\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.939640 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d78ce4-380d-40d0-8f5f-c261776c397c-etc-machine-id\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.940105 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1d78ce4-380d-40d0-8f5f-c261776c397c-logs\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.946735 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.947318 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.947633 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-scripts\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.948246 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data-custom\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:23 crc kubenswrapper[4838]: I1207 10:03:23.963371 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdpxv\" (UniqueName: \"kubernetes.io/projected/c1d78ce4-380d-40d0-8f5f-c261776c397c-kube-api-access-vdpxv\") pod \"manila-api-0\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " pod="openstack/manila-api-0" Dec 07 10:03:24 crc kubenswrapper[4838]: I1207 10:03:24.071440 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 07 10:03:24 crc kubenswrapper[4838]: I1207 10:03:24.281250 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:24 crc kubenswrapper[4838]: I1207 10:03:24.485784 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:24 crc kubenswrapper[4838]: I1207 10:03:24.562497 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-catalog-content\") pod \"d6ca550b-c045-4f06-98de-197e82bc35b1\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " Dec 07 10:03:24 crc kubenswrapper[4838]: I1207 10:03:24.562552 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfnqj\" (UniqueName: \"kubernetes.io/projected/d6ca550b-c045-4f06-98de-197e82bc35b1-kube-api-access-gfnqj\") pod \"d6ca550b-c045-4f06-98de-197e82bc35b1\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " Dec 07 10:03:24 crc kubenswrapper[4838]: I1207 10:03:24.562685 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-utilities\") pod \"d6ca550b-c045-4f06-98de-197e82bc35b1\" (UID: \"d6ca550b-c045-4f06-98de-197e82bc35b1\") " Dec 07 10:03:24 crc kubenswrapper[4838]: I1207 10:03:24.563897 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-utilities" (OuterVolumeSpecName: "utilities") pod "d6ca550b-c045-4f06-98de-197e82bc35b1" (UID: "d6ca550b-c045-4f06-98de-197e82bc35b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.571043 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6ca550b-c045-4f06-98de-197e82bc35b1-kube-api-access-gfnqj" (OuterVolumeSpecName: "kube-api-access-gfnqj") pod "d6ca550b-c045-4f06-98de-197e82bc35b1" (UID: "d6ca550b-c045-4f06-98de-197e82bc35b1"). InnerVolumeSpecName "kube-api-access-gfnqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.645356 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6ca550b-c045-4f06-98de-197e82bc35b1" (UID: "d6ca550b-c045-4f06-98de-197e82bc35b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.665789 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.665829 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6ca550b-c045-4f06-98de-197e82bc35b1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.665840 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfnqj\" (UniqueName: \"kubernetes.io/projected/d6ca550b-c045-4f06-98de-197e82bc35b1-kube-api-access-gfnqj\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:25 crc kubenswrapper[4838]: W1207 10:03:24.679656 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5151e78d_6cd8_4a07_be96_40f44b083861.slice/crio-9447ea061780e8498870ba14e978cf31ca9aed8e201b1742a4437efe36952550 WatchSource:0}: Error finding container 9447ea061780e8498870ba14e978cf31ca9aed8e201b1742a4437efe36952550: Status 404 returned error can't find the container with id 9447ea061780e8498870ba14e978cf31ca9aed8e201b1742a4437efe36952550 Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.682468 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.760836 4838 generic.go:334] "Generic (PLEG): container finished" podID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerID="f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530" exitCode=0 Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.760891 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqfvz" event={"ID":"d6ca550b-c045-4f06-98de-197e82bc35b1","Type":"ContainerDied","Data":"f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530"} Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.760919 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqfvz" event={"ID":"d6ca550b-c045-4f06-98de-197e82bc35b1","Type":"ContainerDied","Data":"6252f72db987d42f4a17b896aff959a777f8b54a3d7689e301e13bfe7c6fcc9e"} Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.760939 4838 scope.go:117] "RemoveContainer" containerID="f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.761580 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqfvz" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.772046 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e0929baa-19bc-4809-95b8-22daa12417f1","Type":"ContainerStarted","Data":"7a9b999bbe513203d9fc8350cf3d9c309ab099540562ebc16ec11f5a6d00fa75"} Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.773930 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7944d6f96c-r7hkg"] Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.775759 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5151e78d-6cd8-4a07-be96-40f44b083861","Type":"ContainerStarted","Data":"9447ea061780e8498870ba14e978cf31ca9aed8e201b1742a4437efe36952550"} Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.826388 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rqfvz"] Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.835549 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rqfvz"] Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.843333 4838 scope.go:117] "RemoveContainer" containerID="6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.887959 4838 scope.go:117] "RemoveContainer" containerID="01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.945346 4838 scope.go:117] "RemoveContainer" containerID="f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530" Dec 07 10:03:25 crc kubenswrapper[4838]: E1207 10:03:24.945643 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530\": container with ID starting with f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530 not found: ID does not exist" containerID="f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.945668 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530"} err="failed to get container status \"f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530\": rpc error: code = NotFound desc = could not find container \"f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530\": container with ID starting with f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530 not found: ID does not exist" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.945687 4838 scope.go:117] "RemoveContainer" containerID="6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77" Dec 07 10:03:25 crc kubenswrapper[4838]: E1207 10:03:24.945906 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77\": container with ID starting with 6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77 not found: ID does not exist" containerID="6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.945926 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77"} err="failed to get container status \"6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77\": rpc error: code = NotFound desc = could not find container \"6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77\": container with ID starting with 6803da4e7627ecc22d0a134e9ff46183a7f1c5a3f300ea8630a0b2f608d39b77 not found: ID does not exist" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.945939 4838 scope.go:117] "RemoveContainer" containerID="01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30" Dec 07 10:03:25 crc kubenswrapper[4838]: E1207 10:03:24.947469 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30\": container with ID starting with 01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30 not found: ID does not exist" containerID="01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.947491 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30"} err="failed to get container status \"01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30\": rpc error: code = NotFound desc = could not find container \"01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30\": container with ID starting with 01d5ef7653b116c54d9546a29f08bdad409b0300b1a41495baf48b40185d3b30 not found: ID does not exist" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:24.947576 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:25 crc kubenswrapper[4838]: E1207 10:03:25.146420 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-conmon-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache]" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:25.633607 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" path="/var/lib/kubelet/pods/d6ca550b-c045-4f06-98de-197e82bc35b1/volumes" Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:25.824368 4838 generic.go:334] "Generic (PLEG): container finished" podID="2e347003-e564-4cf3-b006-9b4b3daf4044" containerID="9fd41a1591916218bd246f1eb9705df58b9510a9c2bce891e8813116b91ab120" exitCode=0 Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:25.824887 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" event={"ID":"2e347003-e564-4cf3-b006-9b4b3daf4044","Type":"ContainerDied","Data":"9fd41a1591916218bd246f1eb9705df58b9510a9c2bce891e8813116b91ab120"} Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:25.824931 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" event={"ID":"2e347003-e564-4cf3-b006-9b4b3daf4044","Type":"ContainerStarted","Data":"6ef6747306296b38a721c9262640872dd7b4629441ed948fc5fbd710be67b607"} Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:25.862441 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c1d78ce4-380d-40d0-8f5f-c261776c397c","Type":"ContainerStarted","Data":"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf"} Dec 07 10:03:25 crc kubenswrapper[4838]: I1207 10:03:25.862498 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c1d78ce4-380d-40d0-8f5f-c261776c397c","Type":"ContainerStarted","Data":"6993d41be18af12dcdea5f0feb99070784cc67f2f14537b8ee9c94689b33c766"} Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.877919 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c1d78ce4-380d-40d0-8f5f-c261776c397c","Type":"ContainerStarted","Data":"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3"} Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.878591 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.890267 4838 generic.go:334] "Generic (PLEG): container finished" podID="5516b801-6088-4d82-be96-ae1efea7283d" containerID="e10c53d56470d945c083cadd2dc46e67f31a74b422934c01ebb00c98b6c97a1f" exitCode=137 Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.890306 4838 generic.go:334] "Generic (PLEG): container finished" podID="5516b801-6088-4d82-be96-ae1efea7283d" containerID="3b2d39e36a4c7493e49f3d690d6dbf8645f0f216d0f81fd814391b65c2bec238" exitCode=137 Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.890353 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868f9df9df-sv4m8" event={"ID":"5516b801-6088-4d82-be96-ae1efea7283d","Type":"ContainerDied","Data":"e10c53d56470d945c083cadd2dc46e67f31a74b422934c01ebb00c98b6c97a1f"} Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.890378 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868f9df9df-sv4m8" event={"ID":"5516b801-6088-4d82-be96-ae1efea7283d","Type":"ContainerDied","Data":"3b2d39e36a4c7493e49f3d690d6dbf8645f0f216d0f81fd814391b65c2bec238"} Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.903235 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" event={"ID":"2e347003-e564-4cf3-b006-9b4b3daf4044","Type":"ContainerStarted","Data":"61dc0df1e3ec5c39ad35655488a92b8c002cdeddb3db0b5de4b9cbd25958a170"} Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.903713 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.913945 4838 generic.go:334] "Generic (PLEG): container finished" podID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerID="38204a3d8e26525ee3d411dd064eeabcc1e8145ef6f662d2fc3fc5f311f00bf2" exitCode=137 Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.913991 4838 generic.go:334] "Generic (PLEG): container finished" podID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerID="6ce0ffbb8cece0655b084ff3ef7b743ad39d153ac9c690af4c46750999125d11" exitCode=137 Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.914109 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d55754b6f-pfkjv" event={"ID":"74a02cf1-7f14-462a-8d5f-8d0df8e8828b","Type":"ContainerDied","Data":"38204a3d8e26525ee3d411dd064eeabcc1e8145ef6f662d2fc3fc5f311f00bf2"} Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.914149 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d55754b6f-pfkjv" event={"ID":"74a02cf1-7f14-462a-8d5f-8d0df8e8828b","Type":"ContainerDied","Data":"6ce0ffbb8cece0655b084ff3ef7b743ad39d153ac9c690af4c46750999125d11"} Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.931496 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.9314340469999998 podStartE2EDuration="3.931434047s" podCreationTimestamp="2025-12-07 10:03:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:03:26.921578451 +0000 UTC m=+3423.628897508" watchObservedRunningTime="2025-12-07 10:03:26.931434047 +0000 UTC m=+3423.638753054" Dec 07 10:03:26 crc kubenswrapper[4838]: I1207 10:03:26.953383 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e0929baa-19bc-4809-95b8-22daa12417f1","Type":"ContainerStarted","Data":"15f92279f55bf64b08b63e516c0252c523be37f5e2f388ada32bcd7964c1229e"} Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.000227 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" podStartSLOduration=4.000203171 podStartE2EDuration="4.000203171s" podCreationTimestamp="2025-12-07 10:03:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:03:26.963038471 +0000 UTC m=+3423.670357498" watchObservedRunningTime="2025-12-07 10:03:27.000203171 +0000 UTC m=+3423.707522188" Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.877512 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.894105 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.895986 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.932635 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-scripts\") pod \"5516b801-6088-4d82-be96-ae1efea7283d\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.932784 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-config-data\") pod \"5516b801-6088-4d82-be96-ae1efea7283d\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.932828 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5516b801-6088-4d82-be96-ae1efea7283d-logs\") pod \"5516b801-6088-4d82-be96-ae1efea7283d\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.932852 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5516b801-6088-4d82-be96-ae1efea7283d-horizon-secret-key\") pod \"5516b801-6088-4d82-be96-ae1efea7283d\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.932976 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbs2d\" (UniqueName: \"kubernetes.io/projected/5516b801-6088-4d82-be96-ae1efea7283d-kube-api-access-jbs2d\") pod \"5516b801-6088-4d82-be96-ae1efea7283d\" (UID: \"5516b801-6088-4d82-be96-ae1efea7283d\") " Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.933591 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5516b801-6088-4d82-be96-ae1efea7283d-logs" (OuterVolumeSpecName: "logs") pod "5516b801-6088-4d82-be96-ae1efea7283d" (UID: "5516b801-6088-4d82-be96-ae1efea7283d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.958001 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5516b801-6088-4d82-be96-ae1efea7283d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5516b801-6088-4d82-be96-ae1efea7283d" (UID: "5516b801-6088-4d82-be96-ae1efea7283d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.987940 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5516b801-6088-4d82-be96-ae1efea7283d-kube-api-access-jbs2d" (OuterVolumeSpecName: "kube-api-access-jbs2d") pod "5516b801-6088-4d82-be96-ae1efea7283d" (UID: "5516b801-6088-4d82-be96-ae1efea7283d"). InnerVolumeSpecName "kube-api-access-jbs2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.999734 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d55754b6f-pfkjv" event={"ID":"74a02cf1-7f14-462a-8d5f-8d0df8e8828b","Type":"ContainerDied","Data":"0103d3fe3d334ded1fb07a74392aff98d5c36a1562007f692c17875d66c82e7c"} Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.999783 4838 scope.go:117] "RemoveContainer" containerID="38204a3d8e26525ee3d411dd064eeabcc1e8145ef6f662d2fc3fc5f311f00bf2" Dec 07 10:03:27 crc kubenswrapper[4838]: I1207 10:03:27.999915 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d55754b6f-pfkjv" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.029099 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868f9df9df-sv4m8" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.029294 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868f9df9df-sv4m8" event={"ID":"5516b801-6088-4d82-be96-ae1efea7283d","Type":"ContainerDied","Data":"98454e31d463719bc79cedcc8d9ebd6e1ae73f5516c13dbaaa8247a459d8e8eb"} Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.042500 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-horizon-secret-key\") pod \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.042672 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-config-data\") pod \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.042848 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-scripts\") pod \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.042873 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mkj2\" (UniqueName: \"kubernetes.io/projected/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-kube-api-access-8mkj2\") pod \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.042914 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-logs\") pod \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\" (UID: \"74a02cf1-7f14-462a-8d5f-8d0df8e8828b\") " Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.043282 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5516b801-6088-4d82-be96-ae1efea7283d-logs\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.043293 4838 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5516b801-6088-4d82-be96-ae1efea7283d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.043302 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbs2d\" (UniqueName: \"kubernetes.io/projected/5516b801-6088-4d82-be96-ae1efea7283d-kube-api-access-jbs2d\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.043590 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-logs" (OuterVolumeSpecName: "logs") pod "74a02cf1-7f14-462a-8d5f-8d0df8e8828b" (UID: "74a02cf1-7f14-462a-8d5f-8d0df8e8828b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.045535 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "74a02cf1-7f14-462a-8d5f-8d0df8e8828b" (UID: "74a02cf1-7f14-462a-8d5f-8d0df8e8828b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.047556 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-config-data" (OuterVolumeSpecName: "config-data") pod "5516b801-6088-4d82-be96-ae1efea7283d" (UID: "5516b801-6088-4d82-be96-ae1efea7283d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.053436 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-kube-api-access-8mkj2" (OuterVolumeSpecName: "kube-api-access-8mkj2") pod "74a02cf1-7f14-462a-8d5f-8d0df8e8828b" (UID: "74a02cf1-7f14-462a-8d5f-8d0df8e8828b"). InnerVolumeSpecName "kube-api-access-8mkj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.133170 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-config-data" (OuterVolumeSpecName: "config-data") pod "74a02cf1-7f14-462a-8d5f-8d0df8e8828b" (UID: "74a02cf1-7f14-462a-8d5f-8d0df8e8828b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.135362 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-scripts" (OuterVolumeSpecName: "scripts") pod "5516b801-6088-4d82-be96-ae1efea7283d" (UID: "5516b801-6088-4d82-be96-ae1efea7283d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.136064 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-scripts" (OuterVolumeSpecName: "scripts") pod "74a02cf1-7f14-462a-8d5f-8d0df8e8828b" (UID: "74a02cf1-7f14-462a-8d5f-8d0df8e8828b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.145952 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.145995 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mkj2\" (UniqueName: \"kubernetes.io/projected/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-kube-api-access-8mkj2\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.146010 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-logs\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.146023 4838 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.146036 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.146048 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a02cf1-7f14-462a-8d5f-8d0df8e8828b-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.146060 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5516b801-6088-4d82-be96-ae1efea7283d-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.378072 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d55754b6f-pfkjv"] Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.400881 4838 scope.go:117] "RemoveContainer" containerID="6ce0ffbb8cece0655b084ff3ef7b743ad39d153ac9c690af4c46750999125d11" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.408652 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d55754b6f-pfkjv"] Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.426927 4838 scope.go:117] "RemoveContainer" containerID="e10c53d56470d945c083cadd2dc46e67f31a74b422934c01ebb00c98b6c97a1f" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.439490 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-868f9df9df-sv4m8"] Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.451314 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-868f9df9df-sv4m8"] Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.676439 4838 scope.go:117] "RemoveContainer" containerID="3b2d39e36a4c7493e49f3d690d6dbf8645f0f216d0f81fd814391b65c2bec238" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.753971 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 10:03:28 crc kubenswrapper[4838]: I1207 10:03:28.998257 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-76f6479d76-n4xcx" podUID="3f6fe9f7-9f77-4da5-823b-b650957c5860" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.241:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 10:03:29 crc kubenswrapper[4838]: I1207 10:03:29.041425 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e0929baa-19bc-4809-95b8-22daa12417f1","Type":"ContainerStarted","Data":"8c91f69f498ca60ec6218ec334bd00bf1cfaa31afea35e33288baa9545199264"} Dec 07 10:03:29 crc kubenswrapper[4838]: I1207 10:03:29.044122 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api-log" containerID="cri-o://2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf" gracePeriod=30 Dec 07 10:03:29 crc kubenswrapper[4838]: I1207 10:03:29.044152 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api" containerID="cri-o://761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3" gracePeriod=30 Dec 07 10:03:29 crc kubenswrapper[4838]: I1207 10:03:29.067293 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=5.631528016 podStartE2EDuration="7.06727505s" podCreationTimestamp="2025-12-07 10:03:22 +0000 UTC" firstStartedPulling="2025-12-07 10:03:24.213144747 +0000 UTC m=+3420.920463764" lastFinishedPulling="2025-12-07 10:03:25.648891781 +0000 UTC m=+3422.356210798" observedRunningTime="2025-12-07 10:03:29.065275104 +0000 UTC m=+3425.772594131" watchObservedRunningTime="2025-12-07 10:03:29.06727505 +0000 UTC m=+3425.774594067" Dec 07 10:03:29 crc kubenswrapper[4838]: I1207 10:03:29.628795 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5516b801-6088-4d82-be96-ae1efea7283d" path="/var/lib/kubelet/pods/5516b801-6088-4d82-be96-ae1efea7283d/volumes" Dec 07 10:03:29 crc kubenswrapper[4838]: I1207 10:03:29.629846 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" path="/var/lib/kubelet/pods/74a02cf1-7f14-462a-8d5f-8d0df8e8828b/volumes" Dec 07 10:03:29 crc kubenswrapper[4838]: I1207 10:03:29.957881 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.058563 4838 generic.go:334] "Generic (PLEG): container finished" podID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerID="761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3" exitCode=0 Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.058593 4838 generic.go:334] "Generic (PLEG): container finished" podID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerID="2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf" exitCode=143 Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.059131 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.059392 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c1d78ce4-380d-40d0-8f5f-c261776c397c","Type":"ContainerDied","Data":"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3"} Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.059449 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c1d78ce4-380d-40d0-8f5f-c261776c397c","Type":"ContainerDied","Data":"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf"} Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.059466 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c1d78ce4-380d-40d0-8f5f-c261776c397c","Type":"ContainerDied","Data":"6993d41be18af12dcdea5f0feb99070784cc67f2f14537b8ee9c94689b33c766"} Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.059485 4838 scope.go:117] "RemoveContainer" containerID="761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.081775 4838 scope.go:117] "RemoveContainer" containerID="2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.113778 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-scripts\") pod \"c1d78ce4-380d-40d0-8f5f-c261776c397c\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.113900 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-combined-ca-bundle\") pod \"c1d78ce4-380d-40d0-8f5f-c261776c397c\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.113969 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdpxv\" (UniqueName: \"kubernetes.io/projected/c1d78ce4-380d-40d0-8f5f-c261776c397c-kube-api-access-vdpxv\") pod \"c1d78ce4-380d-40d0-8f5f-c261776c397c\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.114003 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1d78ce4-380d-40d0-8f5f-c261776c397c-logs\") pod \"c1d78ce4-380d-40d0-8f5f-c261776c397c\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.114030 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d78ce4-380d-40d0-8f5f-c261776c397c-etc-machine-id\") pod \"c1d78ce4-380d-40d0-8f5f-c261776c397c\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.114074 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data-custom\") pod \"c1d78ce4-380d-40d0-8f5f-c261776c397c\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.114105 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data\") pod \"c1d78ce4-380d-40d0-8f5f-c261776c397c\" (UID: \"c1d78ce4-380d-40d0-8f5f-c261776c397c\") " Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.114493 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1d78ce4-380d-40d0-8f5f-c261776c397c-logs" (OuterVolumeSpecName: "logs") pod "c1d78ce4-380d-40d0-8f5f-c261776c397c" (UID: "c1d78ce4-380d-40d0-8f5f-c261776c397c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.114625 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1d78ce4-380d-40d0-8f5f-c261776c397c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c1d78ce4-380d-40d0-8f5f-c261776c397c" (UID: "c1d78ce4-380d-40d0-8f5f-c261776c397c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.115979 4838 scope.go:117] "RemoveContainer" containerID="761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.119255 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3\": container with ID starting with 761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3 not found: ID does not exist" containerID="761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.119297 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3"} err="failed to get container status \"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3\": rpc error: code = NotFound desc = could not find container \"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3\": container with ID starting with 761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3 not found: ID does not exist" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.119320 4838 scope.go:117] "RemoveContainer" containerID="2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.120713 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1d78ce4-380d-40d0-8f5f-c261776c397c-kube-api-access-vdpxv" (OuterVolumeSpecName: "kube-api-access-vdpxv") pod "c1d78ce4-380d-40d0-8f5f-c261776c397c" (UID: "c1d78ce4-380d-40d0-8f5f-c261776c397c"). InnerVolumeSpecName "kube-api-access-vdpxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.121174 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf\": container with ID starting with 2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf not found: ID does not exist" containerID="2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.121211 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf"} err="failed to get container status \"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf\": rpc error: code = NotFound desc = could not find container \"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf\": container with ID starting with 2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf not found: ID does not exist" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.121241 4838 scope.go:117] "RemoveContainer" containerID="761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.122882 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-scripts" (OuterVolumeSpecName: "scripts") pod "c1d78ce4-380d-40d0-8f5f-c261776c397c" (UID: "c1d78ce4-380d-40d0-8f5f-c261776c397c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.129686 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3"} err="failed to get container status \"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3\": rpc error: code = NotFound desc = could not find container \"761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3\": container with ID starting with 761bc709bd1c3fbc9cbf3a6f15989be4e0b7de67215cb4561784225108b195c3 not found: ID does not exist" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.129736 4838 scope.go:117] "RemoveContainer" containerID="2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.129703 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c1d78ce4-380d-40d0-8f5f-c261776c397c" (UID: "c1d78ce4-380d-40d0-8f5f-c261776c397c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.130021 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf"} err="failed to get container status \"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf\": rpc error: code = NotFound desc = could not find container \"2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf\": container with ID starting with 2fe6a1e0fff086f31cbe191060cc2ecaeb7fc071d7c8a5d804f5866565efbcbf not found: ID does not exist" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.167748 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1d78ce4-380d-40d0-8f5f-c261776c397c" (UID: "c1d78ce4-380d-40d0-8f5f-c261776c397c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.184244 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data" (OuterVolumeSpecName: "config-data") pod "c1d78ce4-380d-40d0-8f5f-c261776c397c" (UID: "c1d78ce4-380d-40d0-8f5f-c261776c397c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.216702 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.216733 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdpxv\" (UniqueName: \"kubernetes.io/projected/c1d78ce4-380d-40d0-8f5f-c261776c397c-kube-api-access-vdpxv\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.216746 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1d78ce4-380d-40d0-8f5f-c261776c397c-logs\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.216755 4838 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1d78ce4-380d-40d0-8f5f-c261776c397c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.216763 4838 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.216771 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.216779 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d78ce4-380d-40d0-8f5f-c261776c397c-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.406942 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.432189 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487355 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487810 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487837 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487863 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="extract-utilities" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487870 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="extract-utilities" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487882 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487888 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487903 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487908 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487925 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon-log" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487930 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon-log" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487940 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="extract-content" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487946 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="extract-content" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487953 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="registry-server" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487959 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="registry-server" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487973 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api-log" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487978 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api-log" Dec 07 10:03:30 crc kubenswrapper[4838]: E1207 10:03:30.487988 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon-log" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.487993 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon-log" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.488147 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.488159 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5516b801-6088-4d82-be96-ae1efea7283d" containerName="horizon-log" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.488165 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ca550b-c045-4f06-98de-197e82bc35b1" containerName="registry-server" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.488178 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.488186 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon-log" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.488201 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a02cf1-7f14-462a-8d5f-8d0df8e8828b" containerName="horizon" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.488211 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" containerName="manila-api-log" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.489626 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.495100 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.496383 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.496530 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.500550 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.628978 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-config-data-custom\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629253 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-internal-tls-certs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629405 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79c0afb9-cd0a-4608-87d1-1163282d784a-etc-machine-id\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629486 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-public-tls-certs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629561 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk74w\" (UniqueName: \"kubernetes.io/projected/79c0afb9-cd0a-4608-87d1-1163282d784a-kube-api-access-qk74w\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629639 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-config-data\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629701 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-scripts\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629801 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79c0afb9-cd0a-4608-87d1-1163282d784a-logs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.629882 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731767 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-internal-tls-certs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731867 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79c0afb9-cd0a-4608-87d1-1163282d784a-etc-machine-id\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731889 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-public-tls-certs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731907 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk74w\" (UniqueName: \"kubernetes.io/projected/79c0afb9-cd0a-4608-87d1-1163282d784a-kube-api-access-qk74w\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731940 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-config-data\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731958 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-scripts\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731981 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79c0afb9-cd0a-4608-87d1-1163282d784a-logs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.731997 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.732059 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-config-data-custom\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.732727 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79c0afb9-cd0a-4608-87d1-1163282d784a-etc-machine-id\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.734160 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79c0afb9-cd0a-4608-87d1-1163282d784a-logs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.736447 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-config-data-custom\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.740875 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-internal-tls-certs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.741378 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.741427 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-public-tls-certs\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.753340 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk74w\" (UniqueName: \"kubernetes.io/projected/79c0afb9-cd0a-4608-87d1-1163282d784a-kube-api-access-qk74w\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.754243 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-config-data\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.762284 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79c0afb9-cd0a-4608-87d1-1163282d784a-scripts\") pod \"manila-api-0\" (UID: \"79c0afb9-cd0a-4608-87d1-1163282d784a\") " pod="openstack/manila-api-0" Dec 07 10:03:30 crc kubenswrapper[4838]: I1207 10:03:30.813287 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 07 10:03:31 crc kubenswrapper[4838]: I1207 10:03:31.525347 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 07 10:03:31 crc kubenswrapper[4838]: W1207 10:03:31.536843 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79c0afb9_cd0a_4608_87d1_1163282d784a.slice/crio-c0d3ebc3e6d3ba491d9774a3ff3609af5ebc257844ae9824e4182ce1d42fc7da WatchSource:0}: Error finding container c0d3ebc3e6d3ba491d9774a3ff3609af5ebc257844ae9824e4182ce1d42fc7da: Status 404 returned error can't find the container with id c0d3ebc3e6d3ba491d9774a3ff3609af5ebc257844ae9824e4182ce1d42fc7da Dec 07 10:03:31 crc kubenswrapper[4838]: I1207 10:03:31.633176 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1d78ce4-380d-40d0-8f5f-c261776c397c" path="/var/lib/kubelet/pods/c1d78ce4-380d-40d0-8f5f-c261776c397c/volumes" Dec 07 10:03:32 crc kubenswrapper[4838]: I1207 10:03:32.085146 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"79c0afb9-cd0a-4608-87d1-1163282d784a","Type":"ContainerStarted","Data":"c0d3ebc3e6d3ba491d9774a3ff3609af5ebc257844ae9824e4182ce1d42fc7da"} Dec 07 10:03:33 crc kubenswrapper[4838]: I1207 10:03:33.095022 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"79c0afb9-cd0a-4608-87d1-1163282d784a","Type":"ContainerStarted","Data":"7183b83a366c8ce08c89259f6fb5155cb5eaa55e7de266636f3bb135052da76d"} Dec 07 10:03:33 crc kubenswrapper[4838]: I1207 10:03:33.095359 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"79c0afb9-cd0a-4608-87d1-1163282d784a","Type":"ContainerStarted","Data":"68be10ab7df37ee6ccc6df1337088a28760244f77f8dc4231ed568b8b48b9d5b"} Dec 07 10:03:33 crc kubenswrapper[4838]: I1207 10:03:33.096616 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 07 10:03:33 crc kubenswrapper[4838]: I1207 10:03:33.127498 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.127476048 podStartE2EDuration="3.127476048s" podCreationTimestamp="2025-12-07 10:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:03:33.119856765 +0000 UTC m=+3429.827175782" watchObservedRunningTime="2025-12-07 10:03:33.127476048 +0000 UTC m=+3429.834795065" Dec 07 10:03:33 crc kubenswrapper[4838]: I1207 10:03:33.364520 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 07 10:03:33 crc kubenswrapper[4838]: I1207 10:03:33.908032 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7944d6f96c-r7hkg" Dec 07 10:03:34 crc kubenswrapper[4838]: I1207 10:03:34.001006 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8454874f4c-gkmgf"] Dec 07 10:03:34 crc kubenswrapper[4838]: I1207 10:03:34.003400 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" podUID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerName="dnsmasq-dns" containerID="cri-o://0b3c9353597d233272b320564e6a4203dcd87e3522d5138d7574877b61365b88" gracePeriod=10 Dec 07 10:03:35 crc kubenswrapper[4838]: I1207 10:03:35.137519 4838 generic.go:334] "Generic (PLEG): container finished" podID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerID="0b3c9353597d233272b320564e6a4203dcd87e3522d5138d7574877b61365b88" exitCode=0 Dec 07 10:03:35 crc kubenswrapper[4838]: I1207 10:03:35.137599 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" event={"ID":"8db08fda-8fbf-441f-ac6e-ba4e76ed9480","Type":"ContainerDied","Data":"0b3c9353597d233272b320564e6a4203dcd87e3522d5138d7574877b61365b88"} Dec 07 10:03:35 crc kubenswrapper[4838]: E1207 10:03:35.588228 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-conmon-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache]" Dec 07 10:03:36 crc kubenswrapper[4838]: I1207 10:03:36.269364 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-766695df86-47g96" Dec 07 10:03:36 crc kubenswrapper[4838]: I1207 10:03:36.500370 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.283992 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.294060 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="proxy-httpd" containerID="cri-o://749c49de29181a11711d0234e8033adb4b780ef8ea3126f59efcbd244967798e" gracePeriod=30 Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.294314 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="sg-core" containerID="cri-o://2ff1ac5ffa2469eea370484ef5e53c553c70cdb92d241f06ebc96d0c92f4b517" gracePeriod=30 Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.294374 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-notification-agent" containerID="cri-o://b1eb11aac662b53da64cff6ccda2fc89c6fc85dcedc877014663a7c2491bc73f" gracePeriod=30 Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.294425 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-central-agent" containerID="cri-o://fe6fe567f47b9d35a5f388d8c252e82dc50c4a97b23a78cf6bfae9612098f1b1" gracePeriod=30 Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.422279 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.512486 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-nb\") pod \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.512577 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-dns-svc\") pod \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.512602 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xt5wm\" (UniqueName: \"kubernetes.io/projected/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-kube-api-access-xt5wm\") pod \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.512666 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-sb\") pod \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.512692 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-config\") pod \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.512848 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-openstack-edpm-ipam\") pod \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\" (UID: \"8db08fda-8fbf-441f-ac6e-ba4e76ed9480\") " Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.524394 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-kube-api-access-xt5wm" (OuterVolumeSpecName: "kube-api-access-xt5wm") pod "8db08fda-8fbf-441f-ac6e-ba4e76ed9480" (UID: "8db08fda-8fbf-441f-ac6e-ba4e76ed9480"). InnerVolumeSpecName "kube-api-access-xt5wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.616844 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xt5wm\" (UniqueName: \"kubernetes.io/projected/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-kube-api-access-xt5wm\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.626542 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8db08fda-8fbf-441f-ac6e-ba4e76ed9480" (UID: "8db08fda-8fbf-441f-ac6e-ba4e76ed9480"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.628240 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "8db08fda-8fbf-441f-ac6e-ba4e76ed9480" (UID: "8db08fda-8fbf-441f-ac6e-ba4e76ed9480"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.640793 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8db08fda-8fbf-441f-ac6e-ba4e76ed9480" (UID: "8db08fda-8fbf-441f-ac6e-ba4e76ed9480"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.641708 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8db08fda-8fbf-441f-ac6e-ba4e76ed9480" (UID: "8db08fda-8fbf-441f-ac6e-ba4e76ed9480"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.649350 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-config" (OuterVolumeSpecName: "config") pod "8db08fda-8fbf-441f-ac6e-ba4e76ed9480" (UID: "8db08fda-8fbf-441f-ac6e-ba4e76ed9480"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.718973 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.719010 4838 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.719020 4838 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.719030 4838 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-config\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.719039 4838 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8db08fda-8fbf-441f-ac6e-ba4e76ed9480-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.767656 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-76f6479d76-n4xcx" Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.836208 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-766695df86-47g96"] Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.836625 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon-log" containerID="cri-o://7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0" gracePeriod=30 Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.836933 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" containerID="cri-o://5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5" gracePeriod=30 Dec 07 10:03:38 crc kubenswrapper[4838]: I1207 10:03:38.846218 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.218716 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" event={"ID":"8db08fda-8fbf-441f-ac6e-ba4e76ed9480","Type":"ContainerDied","Data":"12e559f44177754e3d65f30f9147467512a6b5c5119aa99d725c253c5cceb970"} Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.218788 4838 scope.go:117] "RemoveContainer" containerID="0b3c9353597d233272b320564e6a4203dcd87e3522d5138d7574877b61365b88" Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.219006 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454874f4c-gkmgf" Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.244604 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5151e78d-6cd8-4a07-be96-40f44b083861","Type":"ContainerStarted","Data":"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15"} Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.252679 4838 generic.go:334] "Generic (PLEG): container finished" podID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerID="749c49de29181a11711d0234e8033adb4b780ef8ea3126f59efcbd244967798e" exitCode=0 Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.252705 4838 generic.go:334] "Generic (PLEG): container finished" podID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerID="2ff1ac5ffa2469eea370484ef5e53c553c70cdb92d241f06ebc96d0c92f4b517" exitCode=2 Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.252713 4838 generic.go:334] "Generic (PLEG): container finished" podID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerID="fe6fe567f47b9d35a5f388d8c252e82dc50c4a97b23a78cf6bfae9612098f1b1" exitCode=0 Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.252732 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerDied","Data":"749c49de29181a11711d0234e8033adb4b780ef8ea3126f59efcbd244967798e"} Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.252756 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerDied","Data":"2ff1ac5ffa2469eea370484ef5e53c553c70cdb92d241f06ebc96d0c92f4b517"} Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.252766 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerDied","Data":"fe6fe567f47b9d35a5f388d8c252e82dc50c4a97b23a78cf6bfae9612098f1b1"} Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.287515 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8454874f4c-gkmgf"] Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.324652 4838 scope.go:117] "RemoveContainer" containerID="851a9e33e15edd365080edb35f8628de761d087302c83c79ad8eb2d99a020a63" Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.326886 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8454874f4c-gkmgf"] Dec 07 10:03:39 crc kubenswrapper[4838]: I1207 10:03:39.624729 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" path="/var/lib/kubelet/pods/8db08fda-8fbf-441f-ac6e-ba4e76ed9480/volumes" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.265100 4838 generic.go:334] "Generic (PLEG): container finished" podID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerID="b1eb11aac662b53da64cff6ccda2fc89c6fc85dcedc877014663a7c2491bc73f" exitCode=0 Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.265181 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerDied","Data":"b1eb11aac662b53da64cff6ccda2fc89c6fc85dcedc877014663a7c2491bc73f"} Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.267784 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5151e78d-6cd8-4a07-be96-40f44b083861","Type":"ContainerStarted","Data":"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60"} Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.309651 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.929957116 podStartE2EDuration="18.309628551s" podCreationTimestamp="2025-12-07 10:03:22 +0000 UTC" firstStartedPulling="2025-12-07 10:03:24.710191415 +0000 UTC m=+3421.417510432" lastFinishedPulling="2025-12-07 10:03:38.08986285 +0000 UTC m=+3434.797181867" observedRunningTime="2025-12-07 10:03:40.299999042 +0000 UTC m=+3437.007318059" watchObservedRunningTime="2025-12-07 10:03:40.309628551 +0000 UTC m=+3437.016947558" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.602567 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.659264 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-ceilometer-tls-certs\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.660176 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-config-data\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.660293 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-combined-ca-bundle\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.660441 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-log-httpd\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.660606 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-run-httpd\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.660706 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-sg-core-conf-yaml\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.660780 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc9zq\" (UniqueName: \"kubernetes.io/projected/d05b7eba-6424-44d0-bcf9-822d5305f798-kube-api-access-zc9zq\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.660929 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-scripts\") pod \"d05b7eba-6424-44d0-bcf9-822d5305f798\" (UID: \"d05b7eba-6424-44d0-bcf9-822d5305f798\") " Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.661356 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.661537 4838 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.663441 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.673489 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05b7eba-6424-44d0-bcf9-822d5305f798-kube-api-access-zc9zq" (OuterVolumeSpecName: "kube-api-access-zc9zq") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "kube-api-access-zc9zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.680590 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-scripts" (OuterVolumeSpecName: "scripts") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.775185 4838 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05b7eba-6424-44d0-bcf9-822d5305f798-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.775226 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc9zq\" (UniqueName: \"kubernetes.io/projected/d05b7eba-6424-44d0-bcf9-822d5305f798-kube-api-access-zc9zq\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.775238 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.794672 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.799953 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.868432 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.878067 4838 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.878099 4838 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.878109 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.896310 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-config-data" (OuterVolumeSpecName: "config-data") pod "d05b7eba-6424-44d0-bcf9-822d5305f798" (UID: "d05b7eba-6424-44d0-bcf9-822d5305f798"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:40 crc kubenswrapper[4838]: I1207 10:03:40.980347 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05b7eba-6424-44d0-bcf9-822d5305f798-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.202504 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-766695df86-47g96" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.313253 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.314903 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05b7eba-6424-44d0-bcf9-822d5305f798","Type":"ContainerDied","Data":"6a47691762c3038cbb5f49f0c888d76035bacb036fc3dc1174ea94d6c2ab847b"} Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.314971 4838 scope.go:117] "RemoveContainer" containerID="749c49de29181a11711d0234e8033adb4b780ef8ea3126f59efcbd244967798e" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.336580 4838 scope.go:117] "RemoveContainer" containerID="2ff1ac5ffa2469eea370484ef5e53c553c70cdb92d241f06ebc96d0c92f4b517" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.357668 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.373184 4838 scope.go:117] "RemoveContainer" containerID="b1eb11aac662b53da64cff6ccda2fc89c6fc85dcedc877014663a7c2491bc73f" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.383461 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.391866 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 10:03:41 crc kubenswrapper[4838]: E1207 10:03:41.392289 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-notification-agent" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392307 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-notification-agent" Dec 07 10:03:41 crc kubenswrapper[4838]: E1207 10:03:41.392323 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="sg-core" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392330 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="sg-core" Dec 07 10:03:41 crc kubenswrapper[4838]: E1207 10:03:41.392342 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="proxy-httpd" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392348 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="proxy-httpd" Dec 07 10:03:41 crc kubenswrapper[4838]: E1207 10:03:41.392355 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerName="dnsmasq-dns" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392361 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerName="dnsmasq-dns" Dec 07 10:03:41 crc kubenswrapper[4838]: E1207 10:03:41.392398 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerName="init" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392405 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerName="init" Dec 07 10:03:41 crc kubenswrapper[4838]: E1207 10:03:41.392417 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-central-agent" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392422 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-central-agent" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392580 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="sg-core" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392592 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="proxy-httpd" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392603 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-notification-agent" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392619 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8db08fda-8fbf-441f-ac6e-ba4e76ed9480" containerName="dnsmasq-dns" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.392625 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" containerName="ceilometer-central-agent" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.394716 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.400064 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.400275 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.401786 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.407499 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.417109 4838 scope.go:117] "RemoveContainer" containerID="fe6fe567f47b9d35a5f388d8c252e82dc50c4a97b23a78cf6bfae9612098f1b1" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.519229 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-config-data\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.519275 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4997\" (UniqueName: \"kubernetes.io/projected/84d4963b-0485-4254-a707-d621bc87bf6d-kube-api-access-d4997\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.519306 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.519452 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.519542 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d4963b-0485-4254-a707-d621bc87bf6d-log-httpd\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.519756 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.519886 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d4963b-0485-4254-a707-d621bc87bf6d-run-httpd\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.520096 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-scripts\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621606 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-config-data\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621643 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4997\" (UniqueName: \"kubernetes.io/projected/84d4963b-0485-4254-a707-d621bc87bf6d-kube-api-access-d4997\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621670 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621701 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621729 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d4963b-0485-4254-a707-d621bc87bf6d-log-httpd\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621775 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621793 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d4963b-0485-4254-a707-d621bc87bf6d-run-httpd\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.621859 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-scripts\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.624249 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05b7eba-6424-44d0-bcf9-822d5305f798" path="/var/lib/kubelet/pods/d05b7eba-6424-44d0-bcf9-822d5305f798/volumes" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.627583 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d4963b-0485-4254-a707-d621bc87bf6d-run-httpd\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.628114 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-scripts\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.629741 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d4963b-0485-4254-a707-d621bc87bf6d-log-httpd\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.629840 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.630268 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.631051 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.640783 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4963b-0485-4254-a707-d621bc87bf6d-config-data\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.641853 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4997\" (UniqueName: \"kubernetes.io/projected/84d4963b-0485-4254-a707-d621bc87bf6d-kube-api-access-d4997\") pod \"ceilometer-0\" (UID: \"84d4963b-0485-4254-a707-d621bc87bf6d\") " pod="openstack/ceilometer-0" Dec 07 10:03:41 crc kubenswrapper[4838]: I1207 10:03:41.714664 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 10:03:42 crc kubenswrapper[4838]: I1207 10:03:42.240302 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 10:03:42 crc kubenswrapper[4838]: W1207 10:03:42.249678 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d4963b_0485_4254_a707_d621bc87bf6d.slice/crio-336f0aabfc887307795dde58185bdf99e6029f93dcb145355509617a5e1ee36e WatchSource:0}: Error finding container 336f0aabfc887307795dde58185bdf99e6029f93dcb145355509617a5e1ee36e: Status 404 returned error can't find the container with id 336f0aabfc887307795dde58185bdf99e6029f93dcb145355509617a5e1ee36e Dec 07 10:03:42 crc kubenswrapper[4838]: I1207 10:03:42.269353 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 10:03:42 crc kubenswrapper[4838]: I1207 10:03:42.356773 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d4963b-0485-4254-a707-d621bc87bf6d","Type":"ContainerStarted","Data":"336f0aabfc887307795dde58185bdf99e6029f93dcb145355509617a5e1ee36e"} Dec 07 10:03:43 crc kubenswrapper[4838]: I1207 10:03:43.334893 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 07 10:03:43 crc kubenswrapper[4838]: I1207 10:03:43.366265 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d4963b-0485-4254-a707-d621bc87bf6d","Type":"ContainerStarted","Data":"24c0da7990b79f94cbba33a9dc720194aa9844cbf5f19652aa90de22dab396d4"} Dec 07 10:03:44 crc kubenswrapper[4838]: I1207 10:03:44.387887 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d4963b-0485-4254-a707-d621bc87bf6d","Type":"ContainerStarted","Data":"dfd1114d49553b024cc6d3f217d38f3c9e361885b3aa9116b518b66430af3910"} Dec 07 10:03:44 crc kubenswrapper[4838]: I1207 10:03:44.463972 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:34564->10.217.0.240:8443: read: connection reset by peer" Dec 07 10:03:45 crc kubenswrapper[4838]: I1207 10:03:45.274449 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 07 10:03:45 crc kubenswrapper[4838]: I1207 10:03:45.339671 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:45 crc kubenswrapper[4838]: I1207 10:03:45.407515 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d4963b-0485-4254-a707-d621bc87bf6d","Type":"ContainerStarted","Data":"6e88b2620decbaa7177d751f44f1cd384f9e6db6c5868d0dfc27d119de8cbbd9"} Dec 07 10:03:45 crc kubenswrapper[4838]: I1207 10:03:45.409421 4838 generic.go:334] "Generic (PLEG): container finished" podID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerID="5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5" exitCode=0 Dec 07 10:03:45 crc kubenswrapper[4838]: I1207 10:03:45.409555 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766695df86-47g96" event={"ID":"408b9a70-5be1-476e-a026-dbef0631c1d3","Type":"ContainerDied","Data":"5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5"} Dec 07 10:03:45 crc kubenswrapper[4838]: I1207 10:03:45.409620 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="manila-scheduler" containerID="cri-o://15f92279f55bf64b08b63e516c0252c523be37f5e2f388ada32bcd7964c1229e" gracePeriod=30 Dec 07 10:03:45 crc kubenswrapper[4838]: I1207 10:03:45.409657 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="probe" containerID="cri-o://8c91f69f498ca60ec6218ec334bd00bf1cfaa31afea35e33288baa9545199264" gracePeriod=30 Dec 07 10:03:45 crc kubenswrapper[4838]: E1207 10:03:45.887678 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-conmon-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache]" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.436143 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d4963b-0485-4254-a707-d621bc87bf6d","Type":"ContainerStarted","Data":"e76ad648e1053994334194e705d2b8f66e4269ab7b76dca01779ee36bddb59e5"} Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.438355 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.443044 4838 generic.go:334] "Generic (PLEG): container finished" podID="e0929baa-19bc-4809-95b8-22daa12417f1" containerID="8c91f69f498ca60ec6218ec334bd00bf1cfaa31afea35e33288baa9545199264" exitCode=0 Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.443074 4838 generic.go:334] "Generic (PLEG): container finished" podID="e0929baa-19bc-4809-95b8-22daa12417f1" containerID="15f92279f55bf64b08b63e516c0252c523be37f5e2f388ada32bcd7964c1229e" exitCode=0 Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.443116 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e0929baa-19bc-4809-95b8-22daa12417f1","Type":"ContainerDied","Data":"8c91f69f498ca60ec6218ec334bd00bf1cfaa31afea35e33288baa9545199264"} Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.443155 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e0929baa-19bc-4809-95b8-22daa12417f1","Type":"ContainerDied","Data":"15f92279f55bf64b08b63e516c0252c523be37f5e2f388ada32bcd7964c1229e"} Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.470719 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.771127226 podStartE2EDuration="5.470700643s" podCreationTimestamp="2025-12-07 10:03:41 +0000 UTC" firstStartedPulling="2025-12-07 10:03:42.269112839 +0000 UTC m=+3438.976431856" lastFinishedPulling="2025-12-07 10:03:45.968686256 +0000 UTC m=+3442.676005273" observedRunningTime="2025-12-07 10:03:46.454547161 +0000 UTC m=+3443.161866188" watchObservedRunningTime="2025-12-07 10:03:46.470700643 +0000 UTC m=+3443.178019670" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.632427 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.728972 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data\") pod \"e0929baa-19bc-4809-95b8-22daa12417f1\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.729057 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-scripts\") pod \"e0929baa-19bc-4809-95b8-22daa12417f1\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.729541 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0929baa-19bc-4809-95b8-22daa12417f1-etc-machine-id\") pod \"e0929baa-19bc-4809-95b8-22daa12417f1\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.729594 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-combined-ca-bundle\") pod \"e0929baa-19bc-4809-95b8-22daa12417f1\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.729621 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data-custom\") pod \"e0929baa-19bc-4809-95b8-22daa12417f1\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.729648 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0929baa-19bc-4809-95b8-22daa12417f1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e0929baa-19bc-4809-95b8-22daa12417f1" (UID: "e0929baa-19bc-4809-95b8-22daa12417f1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.729661 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4gwg\" (UniqueName: \"kubernetes.io/projected/e0929baa-19bc-4809-95b8-22daa12417f1-kube-api-access-x4gwg\") pod \"e0929baa-19bc-4809-95b8-22daa12417f1\" (UID: \"e0929baa-19bc-4809-95b8-22daa12417f1\") " Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.730372 4838 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0929baa-19bc-4809-95b8-22daa12417f1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.735171 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-scripts" (OuterVolumeSpecName: "scripts") pod "e0929baa-19bc-4809-95b8-22daa12417f1" (UID: "e0929baa-19bc-4809-95b8-22daa12417f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.735207 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0929baa-19bc-4809-95b8-22daa12417f1-kube-api-access-x4gwg" (OuterVolumeSpecName: "kube-api-access-x4gwg") pod "e0929baa-19bc-4809-95b8-22daa12417f1" (UID: "e0929baa-19bc-4809-95b8-22daa12417f1"). InnerVolumeSpecName "kube-api-access-x4gwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.740053 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e0929baa-19bc-4809-95b8-22daa12417f1" (UID: "e0929baa-19bc-4809-95b8-22daa12417f1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.791264 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0929baa-19bc-4809-95b8-22daa12417f1" (UID: "e0929baa-19bc-4809-95b8-22daa12417f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.833111 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.833529 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.833612 4838 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.833686 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4gwg\" (UniqueName: \"kubernetes.io/projected/e0929baa-19bc-4809-95b8-22daa12417f1-kube-api-access-x4gwg\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.849933 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data" (OuterVolumeSpecName: "config-data") pod "e0929baa-19bc-4809-95b8-22daa12417f1" (UID: "e0929baa-19bc-4809-95b8-22daa12417f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:46 crc kubenswrapper[4838]: I1207 10:03:46.935653 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0929baa-19bc-4809-95b8-22daa12417f1-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.457906 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e0929baa-19bc-4809-95b8-22daa12417f1","Type":"ContainerDied","Data":"7a9b999bbe513203d9fc8350cf3d9c309ab099540562ebc16ec11f5a6d00fa75"} Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.457957 4838 scope.go:117] "RemoveContainer" containerID="8c91f69f498ca60ec6218ec334bd00bf1cfaa31afea35e33288baa9545199264" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.458196 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.516106 4838 scope.go:117] "RemoveContainer" containerID="15f92279f55bf64b08b63e516c0252c523be37f5e2f388ada32bcd7964c1229e" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.549974 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.571858 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.580960 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:47 crc kubenswrapper[4838]: E1207 10:03:47.581397 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="manila-scheduler" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.581413 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="manila-scheduler" Dec 07 10:03:47 crc kubenswrapper[4838]: E1207 10:03:47.581443 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="probe" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.581450 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="probe" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.581622 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="manila-scheduler" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.581636 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" containerName="probe" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.582937 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.588724 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.590619 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.624114 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0929baa-19bc-4809-95b8-22daa12417f1" path="/var/lib/kubelet/pods/e0929baa-19bc-4809-95b8-22daa12417f1/volumes" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.647266 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-scripts\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.647315 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-config-data\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.647340 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f404f222-0d07-479e-a0d9-0cb1ce143f3e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.647897 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.647975 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.648146 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcgj2\" (UniqueName: \"kubernetes.io/projected/f404f222-0d07-479e-a0d9-0cb1ce143f3e-kube-api-access-vcgj2\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.749868 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcgj2\" (UniqueName: \"kubernetes.io/projected/f404f222-0d07-479e-a0d9-0cb1ce143f3e-kube-api-access-vcgj2\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.750162 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-scripts\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.750234 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-config-data\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.750301 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f404f222-0d07-479e-a0d9-0cb1ce143f3e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.750362 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f404f222-0d07-479e-a0d9-0cb1ce143f3e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.750518 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.750589 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.754014 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.756870 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.765099 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-scripts\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.766611 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f404f222-0d07-479e-a0d9-0cb1ce143f3e-config-data\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.768318 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcgj2\" (UniqueName: \"kubernetes.io/projected/f404f222-0d07-479e-a0d9-0cb1ce143f3e-kube-api-access-vcgj2\") pod \"manila-scheduler-0\" (UID: \"f404f222-0d07-479e-a0d9-0cb1ce143f3e\") " pod="openstack/manila-scheduler-0" Dec 07 10:03:47 crc kubenswrapper[4838]: I1207 10:03:47.908962 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 07 10:03:48 crc kubenswrapper[4838]: I1207 10:03:48.419213 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 07 10:03:48 crc kubenswrapper[4838]: W1207 10:03:48.423921 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf404f222_0d07_479e_a0d9_0cb1ce143f3e.slice/crio-f7f0558f07ca19c87af0fc9cbe90a7020e270e1fa893240ec89e62bde27d7720 WatchSource:0}: Error finding container f7f0558f07ca19c87af0fc9cbe90a7020e270e1fa893240ec89e62bde27d7720: Status 404 returned error can't find the container with id f7f0558f07ca19c87af0fc9cbe90a7020e270e1fa893240ec89e62bde27d7720 Dec 07 10:03:48 crc kubenswrapper[4838]: I1207 10:03:48.523037 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f404f222-0d07-479e-a0d9-0cb1ce143f3e","Type":"ContainerStarted","Data":"f7f0558f07ca19c87af0fc9cbe90a7020e270e1fa893240ec89e62bde27d7720"} Dec 07 10:03:49 crc kubenswrapper[4838]: I1207 10:03:49.587746 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f404f222-0d07-479e-a0d9-0cb1ce143f3e","Type":"ContainerStarted","Data":"8a2c962266f1af70819a899d3f06b65a2630425e626411486ed37f54d10a9383"} Dec 07 10:03:49 crc kubenswrapper[4838]: I1207 10:03:49.588452 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f404f222-0d07-479e-a0d9-0cb1ce143f3e","Type":"ContainerStarted","Data":"20b49ef59460942dc403042f47c45ffd7f633a1a27018372ff944f87897ad2d4"} Dec 07 10:03:49 crc kubenswrapper[4838]: I1207 10:03:49.623983 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.623957594 podStartE2EDuration="2.623957594s" podCreationTimestamp="2025-12-07 10:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:03:49.618382298 +0000 UTC m=+3446.325701325" watchObservedRunningTime="2025-12-07 10:03:49.623957594 +0000 UTC m=+3446.331276651" Dec 07 10:03:52 crc kubenswrapper[4838]: I1207 10:03:52.280762 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 07 10:03:53 crc kubenswrapper[4838]: I1207 10:03:53.748693 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Dec 07 10:03:54 crc kubenswrapper[4838]: I1207 10:03:54.924596 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 07 10:03:54 crc kubenswrapper[4838]: I1207 10:03:54.993840 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:55 crc kubenswrapper[4838]: I1207 10:03:55.655879 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="manila-share" containerID="cri-o://ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15" gracePeriod=30 Dec 07 10:03:55 crc kubenswrapper[4838]: I1207 10:03:55.655992 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="probe" containerID="cri-o://025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60" gracePeriod=30 Dec 07 10:03:56 crc kubenswrapper[4838]: E1207 10:03:56.119247 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5151e78d_6cd8_4a07_be96_40f44b083861.slice/crio-025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-conmon-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5151e78d_6cd8_4a07_be96_40f44b083861.slice/crio-conmon-ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5151e78d_6cd8_4a07_be96_40f44b083861.slice/crio-conmon-025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60.scope\": RecentStats: unable to find data in memory cache]" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.528829 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569468 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-var-lib-manila\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569530 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569558 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc7zw\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-kube-api-access-bc7zw\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569671 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-etc-machine-id\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569791 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-combined-ca-bundle\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569878 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-scripts\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569936 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-ceph\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.569993 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data-custom\") pod \"5151e78d-6cd8-4a07-be96-40f44b083861\" (UID: \"5151e78d-6cd8-4a07-be96-40f44b083861\") " Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.570974 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.571038 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.582510 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-ceph" (OuterVolumeSpecName: "ceph") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.583062 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-scripts" (OuterVolumeSpecName: "scripts") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.596127 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-kube-api-access-bc7zw" (OuterVolumeSpecName: "kube-api-access-bc7zw") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "kube-api-access-bc7zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.607281 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.668712 4838 generic.go:334] "Generic (PLEG): container finished" podID="5151e78d-6cd8-4a07-be96-40f44b083861" containerID="025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60" exitCode=0 Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.668738 4838 generic.go:334] "Generic (PLEG): container finished" podID="5151e78d-6cd8-4a07-be96-40f44b083861" containerID="ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15" exitCode=1 Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.668757 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5151e78d-6cd8-4a07-be96-40f44b083861","Type":"ContainerDied","Data":"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60"} Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.668782 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5151e78d-6cd8-4a07-be96-40f44b083861","Type":"ContainerDied","Data":"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15"} Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.668793 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5151e78d-6cd8-4a07-be96-40f44b083861","Type":"ContainerDied","Data":"9447ea061780e8498870ba14e978cf31ca9aed8e201b1742a4437efe36952550"} Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.668808 4838 scope.go:117] "RemoveContainer" containerID="025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.668941 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.672197 4838 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.672285 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.672347 4838 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-ceph\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.672398 4838 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.672446 4838 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5151e78d-6cd8-4a07-be96-40f44b083861-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.672496 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc7zw\" (UniqueName: \"kubernetes.io/projected/5151e78d-6cd8-4a07-be96-40f44b083861-kube-api-access-bc7zw\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.678561 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.707764 4838 scope.go:117] "RemoveContainer" containerID="ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.715686 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data" (OuterVolumeSpecName: "config-data") pod "5151e78d-6cd8-4a07-be96-40f44b083861" (UID: "5151e78d-6cd8-4a07-be96-40f44b083861"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.735300 4838 scope.go:117] "RemoveContainer" containerID="025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60" Dec 07 10:03:56 crc kubenswrapper[4838]: E1207 10:03:56.735974 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60\": container with ID starting with 025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60 not found: ID does not exist" containerID="025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.736015 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60"} err="failed to get container status \"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60\": rpc error: code = NotFound desc = could not find container \"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60\": container with ID starting with 025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60 not found: ID does not exist" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.736041 4838 scope.go:117] "RemoveContainer" containerID="ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15" Dec 07 10:03:56 crc kubenswrapper[4838]: E1207 10:03:56.736518 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15\": container with ID starting with ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15 not found: ID does not exist" containerID="ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.736550 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15"} err="failed to get container status \"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15\": rpc error: code = NotFound desc = could not find container \"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15\": container with ID starting with ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15 not found: ID does not exist" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.736574 4838 scope.go:117] "RemoveContainer" containerID="025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.736839 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60"} err="failed to get container status \"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60\": rpc error: code = NotFound desc = could not find container \"025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60\": container with ID starting with 025105d0e90f77a74731a459d0d3bc8ecf742526e847def29b5c1a40814e1f60 not found: ID does not exist" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.736864 4838 scope.go:117] "RemoveContainer" containerID="ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.737162 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15"} err="failed to get container status \"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15\": rpc error: code = NotFound desc = could not find container \"ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15\": container with ID starting with ba666f32bbed2d73ce0c843df82b14ff4fcd7b00150fb36bfae54eda20ab8b15 not found: ID does not exist" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.773261 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.773291 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5151e78d-6cd8-4a07-be96-40f44b083861-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:03:56 crc kubenswrapper[4838]: I1207 10:03:56.997050 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.010286 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.027645 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:57 crc kubenswrapper[4838]: E1207 10:03:57.028107 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="probe" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.028134 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="probe" Dec 07 10:03:57 crc kubenswrapper[4838]: E1207 10:03:57.028158 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="manila-share" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.028166 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="manila-share" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.028417 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="probe" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.028454 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" containerName="manila-share" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.029644 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.033190 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.050860 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.077833 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.077865 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.077886 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5e3f331-938e-4ca9-842e-a580af7035d2-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.077910 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5e3f331-938e-4ca9-842e-a580af7035d2-ceph\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.077956 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b5e3f331-938e-4ca9-842e-a580af7035d2-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.077980 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-config-data\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.077997 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-scripts\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.078024 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2grsl\" (UniqueName: \"kubernetes.io/projected/b5e3f331-938e-4ca9-842e-a580af7035d2-kube-api-access-2grsl\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.179437 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5e3f331-938e-4ca9-842e-a580af7035d2-ceph\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.180504 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b5e3f331-938e-4ca9-842e-a580af7035d2-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.180631 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-config-data\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.180759 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-scripts\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.180897 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2grsl\" (UniqueName: \"kubernetes.io/projected/b5e3f331-938e-4ca9-842e-a580af7035d2-kube-api-access-2grsl\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.181086 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.181181 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.181269 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5e3f331-938e-4ca9-842e-a580af7035d2-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.180635 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b5e3f331-938e-4ca9-842e-a580af7035d2-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.181528 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5e3f331-938e-4ca9-842e-a580af7035d2-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.184502 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b5e3f331-938e-4ca9-842e-a580af7035d2-ceph\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.185312 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-config-data\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.187977 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-scripts\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.191465 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.191585 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e3f331-938e-4ca9-842e-a580af7035d2-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.200168 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2grsl\" (UniqueName: \"kubernetes.io/projected/b5e3f331-938e-4ca9-842e-a580af7035d2-kube-api-access-2grsl\") pod \"manila-share-share1-0\" (UID: \"b5e3f331-938e-4ca9-842e-a580af7035d2\") " pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.347300 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.626657 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5151e78d-6cd8-4a07-be96-40f44b083861" path="/var/lib/kubelet/pods/5151e78d-6cd8-4a07-be96-40f44b083861/volumes" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.909102 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 07 10:03:57 crc kubenswrapper[4838]: I1207 10:03:57.964969 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 07 10:03:58 crc kubenswrapper[4838]: I1207 10:03:58.691309 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b5e3f331-938e-4ca9-842e-a580af7035d2","Type":"ContainerStarted","Data":"7536ed259906f9b82a1e2637a73baa48ae6a52b4884329dea4ef4e3675ce5685"} Dec 07 10:03:58 crc kubenswrapper[4838]: I1207 10:03:58.691685 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b5e3f331-938e-4ca9-842e-a580af7035d2","Type":"ContainerStarted","Data":"99a9fbe7fc35ffe15f994af2a76ed923cbd62269563a3f34263bb5698cdc308e"} Dec 07 10:03:59 crc kubenswrapper[4838]: I1207 10:03:59.701076 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b5e3f331-938e-4ca9-842e-a580af7035d2","Type":"ContainerStarted","Data":"dee10ce761b50cff50b2e0aac8722a04522cc5e38788daf3bd6a14b5c97bd6e7"} Dec 07 10:03:59 crc kubenswrapper[4838]: I1207 10:03:59.728340 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.728322402 podStartE2EDuration="2.728322402s" podCreationTimestamp="2025-12-07 10:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:03:59.720183405 +0000 UTC m=+3456.427502422" watchObservedRunningTime="2025-12-07 10:03:59.728322402 +0000 UTC m=+3456.435641419" Dec 07 10:04:03 crc kubenswrapper[4838]: I1207 10:04:03.751245 4838 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-766695df86-47g96" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Dec 07 10:04:03 crc kubenswrapper[4838]: I1207 10:04:03.751925 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-766695df86-47g96" Dec 07 10:04:06 crc kubenswrapper[4838]: E1207 10:04:06.381028 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-conmon-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache]" Dec 07 10:04:07 crc kubenswrapper[4838]: I1207 10:04:07.347629 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 07 10:04:07 crc kubenswrapper[4838]: I1207 10:04:07.894809 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nsqzl"] Dec 07 10:04:07 crc kubenswrapper[4838]: I1207 10:04:07.897938 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:07 crc kubenswrapper[4838]: I1207 10:04:07.914503 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsqzl"] Dec 07 10:04:07 crc kubenswrapper[4838]: I1207 10:04:07.941789 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-utilities\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:07 crc kubenswrapper[4838]: I1207 10:04:07.941912 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwgrf\" (UniqueName: \"kubernetes.io/projected/c6da4793-4236-47d8-a313-743783f6cf8c-kube-api-access-nwgrf\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:07 crc kubenswrapper[4838]: I1207 10:04:07.941942 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-catalog-content\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.043897 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-utilities\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.044049 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwgrf\" (UniqueName: \"kubernetes.io/projected/c6da4793-4236-47d8-a313-743783f6cf8c-kube-api-access-nwgrf\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.044083 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-catalog-content\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.044510 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-utilities\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.044593 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-catalog-content\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.062893 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwgrf\" (UniqueName: \"kubernetes.io/projected/c6da4793-4236-47d8-a313-743783f6cf8c-kube-api-access-nwgrf\") pod \"community-operators-nsqzl\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.231465 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:08 crc kubenswrapper[4838]: I1207 10:04:08.832274 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsqzl"] Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.327684 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766695df86-47g96" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.471569 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-combined-ca-bundle\") pod \"408b9a70-5be1-476e-a026-dbef0631c1d3\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.471962 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/408b9a70-5be1-476e-a026-dbef0631c1d3-logs\") pod \"408b9a70-5be1-476e-a026-dbef0631c1d3\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.472569 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/408b9a70-5be1-476e-a026-dbef0631c1d3-logs" (OuterVolumeSpecName: "logs") pod "408b9a70-5be1-476e-a026-dbef0631c1d3" (UID: "408b9a70-5be1-476e-a026-dbef0631c1d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.472646 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-secret-key\") pod \"408b9a70-5be1-476e-a026-dbef0631c1d3\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.472684 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7t5j\" (UniqueName: \"kubernetes.io/projected/408b9a70-5be1-476e-a026-dbef0631c1d3-kube-api-access-c7t5j\") pod \"408b9a70-5be1-476e-a026-dbef0631c1d3\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.473183 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-scripts\") pod \"408b9a70-5be1-476e-a026-dbef0631c1d3\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.473232 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-config-data\") pod \"408b9a70-5be1-476e-a026-dbef0631c1d3\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.473292 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-tls-certs\") pod \"408b9a70-5be1-476e-a026-dbef0631c1d3\" (UID: \"408b9a70-5be1-476e-a026-dbef0631c1d3\") " Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.473737 4838 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/408b9a70-5be1-476e-a026-dbef0631c1d3-logs\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.486328 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "408b9a70-5be1-476e-a026-dbef0631c1d3" (UID: "408b9a70-5be1-476e-a026-dbef0631c1d3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.498018 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/408b9a70-5be1-476e-a026-dbef0631c1d3-kube-api-access-c7t5j" (OuterVolumeSpecName: "kube-api-access-c7t5j") pod "408b9a70-5be1-476e-a026-dbef0631c1d3" (UID: "408b9a70-5be1-476e-a026-dbef0631c1d3"). InnerVolumeSpecName "kube-api-access-c7t5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.521534 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-scripts" (OuterVolumeSpecName: "scripts") pod "408b9a70-5be1-476e-a026-dbef0631c1d3" (UID: "408b9a70-5be1-476e-a026-dbef0631c1d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.523901 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-config-data" (OuterVolumeSpecName: "config-data") pod "408b9a70-5be1-476e-a026-dbef0631c1d3" (UID: "408b9a70-5be1-476e-a026-dbef0631c1d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.524079 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "408b9a70-5be1-476e-a026-dbef0631c1d3" (UID: "408b9a70-5be1-476e-a026-dbef0631c1d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.565940 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "408b9a70-5be1-476e-a026-dbef0631c1d3" (UID: "408b9a70-5be1-476e-a026-dbef0631c1d3"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.582214 4838 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.582248 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7t5j\" (UniqueName: \"kubernetes.io/projected/408b9a70-5be1-476e-a026-dbef0631c1d3-kube-api-access-c7t5j\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.582269 4838 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.582279 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/408b9a70-5be1-476e-a026-dbef0631c1d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.582288 4838 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.582296 4838 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b9a70-5be1-476e-a026-dbef0631c1d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.809364 4838 generic.go:334] "Generic (PLEG): container finished" podID="c6da4793-4236-47d8-a313-743783f6cf8c" containerID="de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb" exitCode=0 Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.809673 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsqzl" event={"ID":"c6da4793-4236-47d8-a313-743783f6cf8c","Type":"ContainerDied","Data":"de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb"} Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.809745 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsqzl" event={"ID":"c6da4793-4236-47d8-a313-743783f6cf8c","Type":"ContainerStarted","Data":"1d1c51645c963298192233c04882b5990419fb68eeedecb44a9a9d799dc73b67"} Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.814553 4838 generic.go:334] "Generic (PLEG): container finished" podID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerID="7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0" exitCode=137 Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.814612 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766695df86-47g96" event={"ID":"408b9a70-5be1-476e-a026-dbef0631c1d3","Type":"ContainerDied","Data":"7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0"} Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.814674 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766695df86-47g96" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.814745 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766695df86-47g96" event={"ID":"408b9a70-5be1-476e-a026-dbef0631c1d3","Type":"ContainerDied","Data":"6b0c3f01d610dd79d28c34c005e930b3e3caa5fe96cb21bad719b5ba5b2fa5d3"} Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.814757 4838 scope.go:117] "RemoveContainer" containerID="5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5" Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.868260 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-766695df86-47g96"] Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.877003 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-766695df86-47g96"] Dec 07 10:04:09 crc kubenswrapper[4838]: I1207 10:04:09.952955 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 07 10:04:10 crc kubenswrapper[4838]: I1207 10:04:10.034659 4838 scope.go:117] "RemoveContainer" containerID="7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0" Dec 07 10:04:10 crc kubenswrapper[4838]: I1207 10:04:10.058301 4838 scope.go:117] "RemoveContainer" containerID="5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5" Dec 07 10:04:10 crc kubenswrapper[4838]: E1207 10:04:10.058796 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5\": container with ID starting with 5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5 not found: ID does not exist" containerID="5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5" Dec 07 10:04:10 crc kubenswrapper[4838]: I1207 10:04:10.058956 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5"} err="failed to get container status \"5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5\": rpc error: code = NotFound desc = could not find container \"5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5\": container with ID starting with 5c1900470978ccfe33e6112cbcd736b46a7ed3cdb634b1206e24e12443e627d5 not found: ID does not exist" Dec 07 10:04:10 crc kubenswrapper[4838]: I1207 10:04:10.058991 4838 scope.go:117] "RemoveContainer" containerID="7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0" Dec 07 10:04:10 crc kubenswrapper[4838]: E1207 10:04:10.059401 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0\": container with ID starting with 7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0 not found: ID does not exist" containerID="7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0" Dec 07 10:04:10 crc kubenswrapper[4838]: I1207 10:04:10.059461 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0"} err="failed to get container status \"7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0\": rpc error: code = NotFound desc = could not find container \"7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0\": container with ID starting with 7b1ad65f05878ff1a3489b29a76a2b06a373509175ea966b6528bb8188dbe3c0 not found: ID does not exist" Dec 07 10:04:10 crc kubenswrapper[4838]: I1207 10:04:10.829452 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsqzl" event={"ID":"c6da4793-4236-47d8-a313-743783f6cf8c","Type":"ContainerStarted","Data":"20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4"} Dec 07 10:04:11 crc kubenswrapper[4838]: I1207 10:04:11.632079 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" path="/var/lib/kubelet/pods/408b9a70-5be1-476e-a026-dbef0631c1d3/volumes" Dec 07 10:04:11 crc kubenswrapper[4838]: I1207 10:04:11.798768 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 07 10:04:11 crc kubenswrapper[4838]: I1207 10:04:11.847952 4838 generic.go:334] "Generic (PLEG): container finished" podID="c6da4793-4236-47d8-a313-743783f6cf8c" containerID="20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4" exitCode=0 Dec 07 10:04:11 crc kubenswrapper[4838]: I1207 10:04:11.848009 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsqzl" event={"ID":"c6da4793-4236-47d8-a313-743783f6cf8c","Type":"ContainerDied","Data":"20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4"} Dec 07 10:04:12 crc kubenswrapper[4838]: I1207 10:04:12.861537 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsqzl" event={"ID":"c6da4793-4236-47d8-a313-743783f6cf8c","Type":"ContainerStarted","Data":"310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482"} Dec 07 10:04:12 crc kubenswrapper[4838]: I1207 10:04:12.887446 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nsqzl" podStartSLOduration=3.403077282 podStartE2EDuration="5.887428196s" podCreationTimestamp="2025-12-07 10:04:07 +0000 UTC" firstStartedPulling="2025-12-07 10:04:09.812135917 +0000 UTC m=+3466.519454934" lastFinishedPulling="2025-12-07 10:04:12.296486831 +0000 UTC m=+3469.003805848" observedRunningTime="2025-12-07 10:04:12.882866278 +0000 UTC m=+3469.590185305" watchObservedRunningTime="2025-12-07 10:04:12.887428196 +0000 UTC m=+3469.594747213" Dec 07 10:04:16 crc kubenswrapper[4838]: E1207 10:04:16.605106 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ca550b_c045_4f06_98de_197e82bc35b1.slice/crio-conmon-f51e27a48758f7c7bd4b786f1509a6499f1617e5e1cab84b62cdeac9ac138530.scope\": RecentStats: unable to find data in memory cache]" Dec 07 10:04:18 crc kubenswrapper[4838]: I1207 10:04:18.232355 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:18 crc kubenswrapper[4838]: I1207 10:04:18.232784 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:18 crc kubenswrapper[4838]: I1207 10:04:18.327106 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:18 crc kubenswrapper[4838]: I1207 10:04:18.969073 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:19 crc kubenswrapper[4838]: I1207 10:04:19.047425 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsqzl"] Dec 07 10:04:19 crc kubenswrapper[4838]: I1207 10:04:19.125466 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 07 10:04:20 crc kubenswrapper[4838]: I1207 10:04:20.935840 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nsqzl" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="registry-server" containerID="cri-o://310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482" gracePeriod=2 Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.433651 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.464649 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwgrf\" (UniqueName: \"kubernetes.io/projected/c6da4793-4236-47d8-a313-743783f6cf8c-kube-api-access-nwgrf\") pod \"c6da4793-4236-47d8-a313-743783f6cf8c\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.464700 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-catalog-content\") pod \"c6da4793-4236-47d8-a313-743783f6cf8c\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.465156 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-utilities\") pod \"c6da4793-4236-47d8-a313-743783f6cf8c\" (UID: \"c6da4793-4236-47d8-a313-743783f6cf8c\") " Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.466216 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-utilities" (OuterVolumeSpecName: "utilities") pod "c6da4793-4236-47d8-a313-743783f6cf8c" (UID: "c6da4793-4236-47d8-a313-743783f6cf8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.477028 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6da4793-4236-47d8-a313-743783f6cf8c-kube-api-access-nwgrf" (OuterVolumeSpecName: "kube-api-access-nwgrf") pod "c6da4793-4236-47d8-a313-743783f6cf8c" (UID: "c6da4793-4236-47d8-a313-743783f6cf8c"). InnerVolumeSpecName "kube-api-access-nwgrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.523103 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6da4793-4236-47d8-a313-743783f6cf8c" (UID: "c6da4793-4236-47d8-a313-743783f6cf8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.567239 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.567278 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwgrf\" (UniqueName: \"kubernetes.io/projected/c6da4793-4236-47d8-a313-743783f6cf8c-kube-api-access-nwgrf\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.567293 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da4793-4236-47d8-a313-743783f6cf8c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.955445 4838 generic.go:334] "Generic (PLEG): container finished" podID="c6da4793-4236-47d8-a313-743783f6cf8c" containerID="310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482" exitCode=0 Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.955546 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsqzl" Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.957068 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsqzl" event={"ID":"c6da4793-4236-47d8-a313-743783f6cf8c","Type":"ContainerDied","Data":"310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482"} Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.957282 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsqzl" event={"ID":"c6da4793-4236-47d8-a313-743783f6cf8c","Type":"ContainerDied","Data":"1d1c51645c963298192233c04882b5990419fb68eeedecb44a9a9d799dc73b67"} Dec 07 10:04:21 crc kubenswrapper[4838]: I1207 10:04:21.958666 4838 scope.go:117] "RemoveContainer" containerID="310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.007096 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsqzl"] Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.019471 4838 scope.go:117] "RemoveContainer" containerID="20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.020292 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nsqzl"] Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.053883 4838 scope.go:117] "RemoveContainer" containerID="de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.110204 4838 scope.go:117] "RemoveContainer" containerID="310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482" Dec 07 10:04:22 crc kubenswrapper[4838]: E1207 10:04:22.111351 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482\": container with ID starting with 310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482 not found: ID does not exist" containerID="310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.111408 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482"} err="failed to get container status \"310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482\": rpc error: code = NotFound desc = could not find container \"310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482\": container with ID starting with 310fc32f80616d36d74078c614a1ce19ea998fc9175110e6fcbc1e7e63607482 not found: ID does not exist" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.111440 4838 scope.go:117] "RemoveContainer" containerID="20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4" Dec 07 10:04:22 crc kubenswrapper[4838]: E1207 10:04:22.111793 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4\": container with ID starting with 20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4 not found: ID does not exist" containerID="20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.111874 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4"} err="failed to get container status \"20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4\": rpc error: code = NotFound desc = could not find container \"20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4\": container with ID starting with 20a6d847740b64b6459cdeb8f305c602069bc6d9f9c88bb410660259c5fba3a4 not found: ID does not exist" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.111912 4838 scope.go:117] "RemoveContainer" containerID="de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb" Dec 07 10:04:22 crc kubenswrapper[4838]: E1207 10:04:22.112204 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb\": container with ID starting with de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb not found: ID does not exist" containerID="de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb" Dec 07 10:04:22 crc kubenswrapper[4838]: I1207 10:04:22.112245 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb"} err="failed to get container status \"de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb\": rpc error: code = NotFound desc = could not find container \"de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb\": container with ID starting with de9d9b94de47e275c3c4ae21944b37c954fa99e1c3bc4ff4031e7cc21e4436fb not found: ID does not exist" Dec 07 10:04:23 crc kubenswrapper[4838]: I1207 10:04:23.632510 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" path="/var/lib/kubelet/pods/c6da4793-4236-47d8-a313-743783f6cf8c/volumes" Dec 07 10:04:58 crc kubenswrapper[4838]: E1207 10:04:58.629740 4838 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.73:36856->38.102.83.73:33371: read tcp 38.102.83.73:36856->38.102.83.73:33371: read: connection reset by peer Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.557289 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 07 10:05:23 crc kubenswrapper[4838]: E1207 10:05:23.558390 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon-log" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558413 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon-log" Dec 07 10:05:23 crc kubenswrapper[4838]: E1207 10:05:23.558433 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="extract-utilities" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558444 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="extract-utilities" Dec 07 10:05:23 crc kubenswrapper[4838]: E1207 10:05:23.558470 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558481 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" Dec 07 10:05:23 crc kubenswrapper[4838]: E1207 10:05:23.558498 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="extract-content" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558508 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="extract-content" Dec 07 10:05:23 crc kubenswrapper[4838]: E1207 10:05:23.558528 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="registry-server" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558537 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="registry-server" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558896 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558942 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6da4793-4236-47d8-a313-743783f6cf8c" containerName="registry-server" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.558964 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="408b9a70-5be1-476e-a026-dbef0631c1d3" containerName="horizon-log" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.559937 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.564329 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.564538 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.564551 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pcr24" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.564593 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.573974 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632539 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632599 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632658 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-config-data\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632675 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632706 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632721 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632739 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtb2k\" (UniqueName: \"kubernetes.io/projected/a430a791-ad62-4246-8422-6bae0b6ce772-kube-api-access-mtb2k\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632786 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.632812 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735058 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735136 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-config-data\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735157 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735194 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735217 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735244 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtb2k\" (UniqueName: \"kubernetes.io/projected/a430a791-ad62-4246-8422-6bae0b6ce772-kube-api-access-mtb2k\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735295 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735332 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735431 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735563 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.735657 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.737140 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.737243 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.737390 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.737953 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.742306 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.750885 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.751877 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-config-data\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.752222 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.768734 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtb2k\" (UniqueName: \"kubernetes.io/projected/a430a791-ad62-4246-8422-6bae0b6ce772-kube-api-access-mtb2k\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.769161 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " pod="openstack/tempest-tests-tempest" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.892169 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pcr24" Dec 07 10:05:23 crc kubenswrapper[4838]: I1207 10:05:23.937156 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 10:05:24 crc kubenswrapper[4838]: I1207 10:05:24.464638 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 07 10:05:24 crc kubenswrapper[4838]: I1207 10:05:24.492881 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:05:24 crc kubenswrapper[4838]: I1207 10:05:24.492930 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:05:24 crc kubenswrapper[4838]: I1207 10:05:24.609396 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a430a791-ad62-4246-8422-6bae0b6ce772","Type":"ContainerStarted","Data":"8f3b31953ba8c0eb4c530a1df4c35a35649b0005e492511bec032ae8eac4acbc"} Dec 07 10:05:54 crc kubenswrapper[4838]: I1207 10:05:54.492708 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:05:54 crc kubenswrapper[4838]: I1207 10:05:54.493307 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:06:04 crc kubenswrapper[4838]: E1207 10:06:04.553761 4838 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 07 10:06:04 crc kubenswrapper[4838]: E1207 10:06:04.555140 4838 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mtb2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(a430a791-ad62-4246-8422-6bae0b6ce772): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 10:06:04 crc kubenswrapper[4838]: E1207 10:06:04.556373 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="a430a791-ad62-4246-8422-6bae0b6ce772" Dec 07 10:06:05 crc kubenswrapper[4838]: E1207 10:06:05.014188 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="a430a791-ad62-4246-8422-6bae0b6ce772" Dec 07 10:06:20 crc kubenswrapper[4838]: I1207 10:06:20.060135 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 07 10:06:22 crc kubenswrapper[4838]: I1207 10:06:22.210701 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a430a791-ad62-4246-8422-6bae0b6ce772","Type":"ContainerStarted","Data":"73700bfabf6d58d2ddafd40f5b882a22f9a570bb72e32c37029a1a76af6b9805"} Dec 07 10:06:22 crc kubenswrapper[4838]: I1207 10:06:22.242190 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.649527141 podStartE2EDuration="1m0.242167865s" podCreationTimestamp="2025-12-07 10:05:22 +0000 UTC" firstStartedPulling="2025-12-07 10:05:24.462854048 +0000 UTC m=+3541.170173065" lastFinishedPulling="2025-12-07 10:06:20.055494762 +0000 UTC m=+3596.762813789" observedRunningTime="2025-12-07 10:06:22.240983554 +0000 UTC m=+3598.948302581" watchObservedRunningTime="2025-12-07 10:06:22.242167865 +0000 UTC m=+3598.949486892" Dec 07 10:06:24 crc kubenswrapper[4838]: I1207 10:06:24.493675 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:06:24 crc kubenswrapper[4838]: I1207 10:06:24.494824 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:06:24 crc kubenswrapper[4838]: I1207 10:06:24.494930 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 10:06:24 crc kubenswrapper[4838]: I1207 10:06:24.495655 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 10:06:24 crc kubenswrapper[4838]: I1207 10:06:24.495772 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" gracePeriod=600 Dec 07 10:06:24 crc kubenswrapper[4838]: E1207 10:06:24.619544 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:06:25 crc kubenswrapper[4838]: I1207 10:06:25.259277 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" exitCode=0 Dec 07 10:06:25 crc kubenswrapper[4838]: I1207 10:06:25.259328 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3"} Dec 07 10:06:25 crc kubenswrapper[4838]: I1207 10:06:25.259378 4838 scope.go:117] "RemoveContainer" containerID="9f7e06fe287c32a6d89d0c626a75199a006e5e2536c384a448e44b3eb61037c0" Dec 07 10:06:25 crc kubenswrapper[4838]: I1207 10:06:25.260382 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:06:25 crc kubenswrapper[4838]: E1207 10:06:25.261000 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:06:39 crc kubenswrapper[4838]: I1207 10:06:39.615838 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:06:39 crc kubenswrapper[4838]: E1207 10:06:39.616655 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:06:52 crc kubenswrapper[4838]: I1207 10:06:52.615017 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:06:52 crc kubenswrapper[4838]: E1207 10:06:52.615755 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:07:06 crc kubenswrapper[4838]: I1207 10:07:06.614692 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:07:06 crc kubenswrapper[4838]: E1207 10:07:06.615648 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:07:17 crc kubenswrapper[4838]: I1207 10:07:17.614601 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:07:17 crc kubenswrapper[4838]: E1207 10:07:17.615297 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:07:31 crc kubenswrapper[4838]: I1207 10:07:31.614373 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:07:31 crc kubenswrapper[4838]: E1207 10:07:31.615370 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:07:42 crc kubenswrapper[4838]: I1207 10:07:42.616683 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:07:42 crc kubenswrapper[4838]: E1207 10:07:42.617792 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:07:53 crc kubenswrapper[4838]: I1207 10:07:53.630653 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:07:53 crc kubenswrapper[4838]: E1207 10:07:53.632198 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:08:08 crc kubenswrapper[4838]: I1207 10:08:08.615429 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:08:08 crc kubenswrapper[4838]: E1207 10:08:08.616495 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:08:23 crc kubenswrapper[4838]: I1207 10:08:23.633775 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:08:23 crc kubenswrapper[4838]: E1207 10:08:23.634406 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:08:34 crc kubenswrapper[4838]: I1207 10:08:34.614830 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:08:34 crc kubenswrapper[4838]: E1207 10:08:34.615486 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:08:49 crc kubenswrapper[4838]: I1207 10:08:49.614947 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:08:49 crc kubenswrapper[4838]: E1207 10:08:49.615884 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:09:00 crc kubenswrapper[4838]: I1207 10:09:00.615014 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:09:00 crc kubenswrapper[4838]: E1207 10:09:00.615958 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:09:11 crc kubenswrapper[4838]: I1207 10:09:11.614227 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:09:11 crc kubenswrapper[4838]: E1207 10:09:11.615043 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:09:26 crc kubenswrapper[4838]: I1207 10:09:26.614494 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:09:26 crc kubenswrapper[4838]: E1207 10:09:26.615334 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:09:40 crc kubenswrapper[4838]: I1207 10:09:40.615586 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:09:40 crc kubenswrapper[4838]: E1207 10:09:40.618455 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:09:53 crc kubenswrapper[4838]: I1207 10:09:53.179762 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:09:53 crc kubenswrapper[4838]: E1207 10:09:53.180812 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:10:08 crc kubenswrapper[4838]: I1207 10:10:08.614696 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:10:08 crc kubenswrapper[4838]: E1207 10:10:08.615341 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:10:19 crc kubenswrapper[4838]: I1207 10:10:19.616013 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:10:19 crc kubenswrapper[4838]: E1207 10:10:19.616995 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:10:32 crc kubenswrapper[4838]: I1207 10:10:32.615937 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:10:32 crc kubenswrapper[4838]: E1207 10:10:32.616683 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:10:43 crc kubenswrapper[4838]: I1207 10:10:43.622612 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:10:43 crc kubenswrapper[4838]: E1207 10:10:43.623331 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:10:57 crc kubenswrapper[4838]: I1207 10:10:57.614792 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:10:57 crc kubenswrapper[4838]: E1207 10:10:57.616017 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:11:09 crc kubenswrapper[4838]: I1207 10:11:09.616441 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:11:09 crc kubenswrapper[4838]: E1207 10:11:09.617240 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:11:23 crc kubenswrapper[4838]: I1207 10:11:23.628585 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:11:23 crc kubenswrapper[4838]: E1207 10:11:23.629325 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:11:36 crc kubenswrapper[4838]: I1207 10:11:36.614731 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:11:37 crc kubenswrapper[4838]: I1207 10:11:37.266553 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"8c12cece5226220a2db5c267c84bcf3318cdd31d03417d7b5d04de254f9f1a93"} Dec 07 10:12:25 crc kubenswrapper[4838]: I1207 10:12:25.863909 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rv7l8"] Dec 07 10:12:25 crc kubenswrapper[4838]: I1207 10:12:25.866305 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:25 crc kubenswrapper[4838]: I1207 10:12:25.939254 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6jmh\" (UniqueName: \"kubernetes.io/projected/934898b1-e68f-4b10-9679-805ce7c40c02-kube-api-access-r6jmh\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:25 crc kubenswrapper[4838]: I1207 10:12:25.939455 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-catalog-content\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:25 crc kubenswrapper[4838]: I1207 10:12:25.939647 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-utilities\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:25 crc kubenswrapper[4838]: I1207 10:12:25.961074 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rv7l8"] Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.055708 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6jmh\" (UniqueName: \"kubernetes.io/projected/934898b1-e68f-4b10-9679-805ce7c40c02-kube-api-access-r6jmh\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.055793 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-catalog-content\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.055961 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-utilities\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.056488 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-utilities\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.056875 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-catalog-content\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.080707 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6jmh\" (UniqueName: \"kubernetes.io/projected/934898b1-e68f-4b10-9679-805ce7c40c02-kube-api-access-r6jmh\") pod \"redhat-operators-rv7l8\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.183581 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:26 crc kubenswrapper[4838]: I1207 10:12:26.903010 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rv7l8"] Dec 07 10:12:27 crc kubenswrapper[4838]: I1207 10:12:27.726047 4838 generic.go:334] "Generic (PLEG): container finished" podID="934898b1-e68f-4b10-9679-805ce7c40c02" containerID="6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2" exitCode=0 Dec 07 10:12:27 crc kubenswrapper[4838]: I1207 10:12:27.727425 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rv7l8" event={"ID":"934898b1-e68f-4b10-9679-805ce7c40c02","Type":"ContainerDied","Data":"6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2"} Dec 07 10:12:27 crc kubenswrapper[4838]: I1207 10:12:27.727453 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rv7l8" event={"ID":"934898b1-e68f-4b10-9679-805ce7c40c02","Type":"ContainerStarted","Data":"a279cd4d05ada4e6b43b109536066db0f824fef7a08a458c7c93e60e449a91fe"} Dec 07 10:12:27 crc kubenswrapper[4838]: I1207 10:12:27.728296 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 10:12:28 crc kubenswrapper[4838]: I1207 10:12:28.739753 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rv7l8" event={"ID":"934898b1-e68f-4b10-9679-805ce7c40c02","Type":"ContainerStarted","Data":"68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f"} Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.061903 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mk7kb"] Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.064833 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.074178 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mk7kb"] Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.114652 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-utilities\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.114730 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-catalog-content\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.114839 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98m86\" (UniqueName: \"kubernetes.io/projected/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-kube-api-access-98m86\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.216101 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98m86\" (UniqueName: \"kubernetes.io/projected/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-kube-api-access-98m86\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.216516 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-utilities\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.216593 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-catalog-content\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.219679 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-utilities\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.219921 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-catalog-content\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.245523 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98m86\" (UniqueName: \"kubernetes.io/projected/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-kube-api-access-98m86\") pod \"redhat-marketplace-mk7kb\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:29 crc kubenswrapper[4838]: I1207 10:12:29.415059 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:30 crc kubenswrapper[4838]: I1207 10:12:30.265950 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mk7kb"] Dec 07 10:12:30 crc kubenswrapper[4838]: W1207 10:12:30.294903 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92deb8e6_c5a8_44da_b1a4_e02ab7520b74.slice/crio-e2f3251cd49db2ced88adcceb1c80c955e7a85ce24d94e10fb03f68b2e8cf67c WatchSource:0}: Error finding container e2f3251cd49db2ced88adcceb1c80c955e7a85ce24d94e10fb03f68b2e8cf67c: Status 404 returned error can't find the container with id e2f3251cd49db2ced88adcceb1c80c955e7a85ce24d94e10fb03f68b2e8cf67c Dec 07 10:12:30 crc kubenswrapper[4838]: I1207 10:12:30.770468 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mk7kb" event={"ID":"92deb8e6-c5a8-44da-b1a4-e02ab7520b74","Type":"ContainerStarted","Data":"fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f"} Dec 07 10:12:30 crc kubenswrapper[4838]: I1207 10:12:30.771103 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mk7kb" event={"ID":"92deb8e6-c5a8-44da-b1a4-e02ab7520b74","Type":"ContainerStarted","Data":"e2f3251cd49db2ced88adcceb1c80c955e7a85ce24d94e10fb03f68b2e8cf67c"} Dec 07 10:12:31 crc kubenswrapper[4838]: I1207 10:12:31.779300 4838 generic.go:334] "Generic (PLEG): container finished" podID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerID="fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f" exitCode=0 Dec 07 10:12:31 crc kubenswrapper[4838]: I1207 10:12:31.779374 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mk7kb" event={"ID":"92deb8e6-c5a8-44da-b1a4-e02ab7520b74","Type":"ContainerDied","Data":"fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f"} Dec 07 10:12:31 crc kubenswrapper[4838]: I1207 10:12:31.781430 4838 generic.go:334] "Generic (PLEG): container finished" podID="934898b1-e68f-4b10-9679-805ce7c40c02" containerID="68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f" exitCode=0 Dec 07 10:12:31 crc kubenswrapper[4838]: I1207 10:12:31.781465 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rv7l8" event={"ID":"934898b1-e68f-4b10-9679-805ce7c40c02","Type":"ContainerDied","Data":"68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f"} Dec 07 10:12:32 crc kubenswrapper[4838]: I1207 10:12:32.791008 4838 generic.go:334] "Generic (PLEG): container finished" podID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerID="afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177" exitCode=0 Dec 07 10:12:32 crc kubenswrapper[4838]: I1207 10:12:32.791109 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mk7kb" event={"ID":"92deb8e6-c5a8-44da-b1a4-e02ab7520b74","Type":"ContainerDied","Data":"afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177"} Dec 07 10:12:32 crc kubenswrapper[4838]: I1207 10:12:32.795553 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rv7l8" event={"ID":"934898b1-e68f-4b10-9679-805ce7c40c02","Type":"ContainerStarted","Data":"358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0"} Dec 07 10:12:32 crc kubenswrapper[4838]: I1207 10:12:32.844660 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rv7l8" podStartSLOduration=3.412322354 podStartE2EDuration="7.844634931s" podCreationTimestamp="2025-12-07 10:12:25 +0000 UTC" firstStartedPulling="2025-12-07 10:12:27.727919434 +0000 UTC m=+3964.435238471" lastFinishedPulling="2025-12-07 10:12:32.160232031 +0000 UTC m=+3968.867551048" observedRunningTime="2025-12-07 10:12:32.826437626 +0000 UTC m=+3969.533756643" watchObservedRunningTime="2025-12-07 10:12:32.844634931 +0000 UTC m=+3969.551953958" Dec 07 10:12:33 crc kubenswrapper[4838]: I1207 10:12:33.804260 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mk7kb" event={"ID":"92deb8e6-c5a8-44da-b1a4-e02ab7520b74","Type":"ContainerStarted","Data":"f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf"} Dec 07 10:12:33 crc kubenswrapper[4838]: I1207 10:12:33.848726 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mk7kb" podStartSLOduration=2.412013464 podStartE2EDuration="4.848703999s" podCreationTimestamp="2025-12-07 10:12:29 +0000 UTC" firstStartedPulling="2025-12-07 10:12:30.776258711 +0000 UTC m=+3967.483577738" lastFinishedPulling="2025-12-07 10:12:33.212949246 +0000 UTC m=+3969.920268273" observedRunningTime="2025-12-07 10:12:33.842183625 +0000 UTC m=+3970.549502652" watchObservedRunningTime="2025-12-07 10:12:33.848703999 +0000 UTC m=+3970.556023016" Dec 07 10:12:35 crc kubenswrapper[4838]: I1207 10:12:35.063105 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-7e54-account-create-update-c6vjz"] Dec 07 10:12:35 crc kubenswrapper[4838]: I1207 10:12:35.070995 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-7e54-account-create-update-c6vjz"] Dec 07 10:12:35 crc kubenswrapper[4838]: I1207 10:12:35.627225 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92cc0e9b-6cd2-4a17-9e61-95931c154279" path="/var/lib/kubelet/pods/92cc0e9b-6cd2-4a17-9e61-95931c154279/volumes" Dec 07 10:12:36 crc kubenswrapper[4838]: I1207 10:12:36.068181 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-gzs9g"] Dec 07 10:12:36 crc kubenswrapper[4838]: I1207 10:12:36.078475 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-gzs9g"] Dec 07 10:12:36 crc kubenswrapper[4838]: I1207 10:12:36.184623 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:36 crc kubenswrapper[4838]: I1207 10:12:36.184682 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:37 crc kubenswrapper[4838]: I1207 10:12:37.245523 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rv7l8" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="registry-server" probeResult="failure" output=< Dec 07 10:12:37 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 10:12:37 crc kubenswrapper[4838]: > Dec 07 10:12:37 crc kubenswrapper[4838]: I1207 10:12:37.626930 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="521c679e-583d-4ee9-827e-bb3b4a1e7e1e" path="/var/lib/kubelet/pods/521c679e-583d-4ee9-827e-bb3b4a1e7e1e/volumes" Dec 07 10:12:39 crc kubenswrapper[4838]: I1207 10:12:39.415193 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:39 crc kubenswrapper[4838]: I1207 10:12:39.415443 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:39 crc kubenswrapper[4838]: I1207 10:12:39.465286 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:39 crc kubenswrapper[4838]: I1207 10:12:39.910684 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:39 crc kubenswrapper[4838]: I1207 10:12:39.954920 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mk7kb"] Dec 07 10:12:41 crc kubenswrapper[4838]: I1207 10:12:41.870198 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mk7kb" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="registry-server" containerID="cri-o://f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf" gracePeriod=2 Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.450333 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.499402 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-utilities\") pod \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.499879 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98m86\" (UniqueName: \"kubernetes.io/projected/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-kube-api-access-98m86\") pod \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.499983 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-catalog-content\") pod \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\" (UID: \"92deb8e6-c5a8-44da-b1a4-e02ab7520b74\") " Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.500873 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-utilities" (OuterVolumeSpecName: "utilities") pod "92deb8e6-c5a8-44da-b1a4-e02ab7520b74" (UID: "92deb8e6-c5a8-44da-b1a4-e02ab7520b74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.526711 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-kube-api-access-98m86" (OuterVolumeSpecName: "kube-api-access-98m86") pod "92deb8e6-c5a8-44da-b1a4-e02ab7520b74" (UID: "92deb8e6-c5a8-44da-b1a4-e02ab7520b74"). InnerVolumeSpecName "kube-api-access-98m86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.535232 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92deb8e6-c5a8-44da-b1a4-e02ab7520b74" (UID: "92deb8e6-c5a8-44da-b1a4-e02ab7520b74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.603475 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98m86\" (UniqueName: \"kubernetes.io/projected/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-kube-api-access-98m86\") on node \"crc\" DevicePath \"\"" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.603510 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.603519 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92deb8e6-c5a8-44da-b1a4-e02ab7520b74-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.880583 4838 generic.go:334] "Generic (PLEG): container finished" podID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerID="f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf" exitCode=0 Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.880630 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mk7kb" event={"ID":"92deb8e6-c5a8-44da-b1a4-e02ab7520b74","Type":"ContainerDied","Data":"f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf"} Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.880649 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mk7kb" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.880659 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mk7kb" event={"ID":"92deb8e6-c5a8-44da-b1a4-e02ab7520b74","Type":"ContainerDied","Data":"e2f3251cd49db2ced88adcceb1c80c955e7a85ce24d94e10fb03f68b2e8cf67c"} Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.880677 4838 scope.go:117] "RemoveContainer" containerID="f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.899902 4838 scope.go:117] "RemoveContainer" containerID="afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.922445 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mk7kb"] Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.929636 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mk7kb"] Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.950991 4838 scope.go:117] "RemoveContainer" containerID="fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.988875 4838 scope.go:117] "RemoveContainer" containerID="f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf" Dec 07 10:12:42 crc kubenswrapper[4838]: E1207 10:12:42.990666 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf\": container with ID starting with f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf not found: ID does not exist" containerID="f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.990698 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf"} err="failed to get container status \"f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf\": rpc error: code = NotFound desc = could not find container \"f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf\": container with ID starting with f746b5e51045ce12b7fc057b4a892cb1941e4befef17cb0a277052cb273678bf not found: ID does not exist" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.990720 4838 scope.go:117] "RemoveContainer" containerID="afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177" Dec 07 10:12:42 crc kubenswrapper[4838]: E1207 10:12:42.991116 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177\": container with ID starting with afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177 not found: ID does not exist" containerID="afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.992062 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177"} err="failed to get container status \"afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177\": rpc error: code = NotFound desc = could not find container \"afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177\": container with ID starting with afd1f1a80b1c998f71c3c025f7033465b6c1751858288b9e2d556a98ed120177 not found: ID does not exist" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.992088 4838 scope.go:117] "RemoveContainer" containerID="fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f" Dec 07 10:12:42 crc kubenswrapper[4838]: E1207 10:12:42.992732 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f\": container with ID starting with fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f not found: ID does not exist" containerID="fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f" Dec 07 10:12:42 crc kubenswrapper[4838]: I1207 10:12:42.994206 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f"} err="failed to get container status \"fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f\": rpc error: code = NotFound desc = could not find container \"fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f\": container with ID starting with fdc5ce336d5b6f86e019e734df9e6486f70ca5c9eacee7704e6502e66e42752f not found: ID does not exist" Dec 07 10:12:43 crc kubenswrapper[4838]: I1207 10:12:43.627033 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" path="/var/lib/kubelet/pods/92deb8e6-c5a8-44da-b1a4-e02ab7520b74/volumes" Dec 07 10:12:46 crc kubenswrapper[4838]: I1207 10:12:46.231744 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:46 crc kubenswrapper[4838]: I1207 10:12:46.287070 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:47 crc kubenswrapper[4838]: I1207 10:12:47.134281 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rv7l8"] Dec 07 10:12:47 crc kubenswrapper[4838]: I1207 10:12:47.928155 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rv7l8" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="registry-server" containerID="cri-o://358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0" gracePeriod=2 Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.554397 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.646429 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-utilities\") pod \"934898b1-e68f-4b10-9679-805ce7c40c02\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.646538 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-catalog-content\") pod \"934898b1-e68f-4b10-9679-805ce7c40c02\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.646577 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6jmh\" (UniqueName: \"kubernetes.io/projected/934898b1-e68f-4b10-9679-805ce7c40c02-kube-api-access-r6jmh\") pod \"934898b1-e68f-4b10-9679-805ce7c40c02\" (UID: \"934898b1-e68f-4b10-9679-805ce7c40c02\") " Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.647522 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-utilities" (OuterVolumeSpecName: "utilities") pod "934898b1-e68f-4b10-9679-805ce7c40c02" (UID: "934898b1-e68f-4b10-9679-805ce7c40c02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.652493 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/934898b1-e68f-4b10-9679-805ce7c40c02-kube-api-access-r6jmh" (OuterVolumeSpecName: "kube-api-access-r6jmh") pod "934898b1-e68f-4b10-9679-805ce7c40c02" (UID: "934898b1-e68f-4b10-9679-805ce7c40c02"). InnerVolumeSpecName "kube-api-access-r6jmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.747259 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "934898b1-e68f-4b10-9679-805ce7c40c02" (UID: "934898b1-e68f-4b10-9679-805ce7c40c02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.748691 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6jmh\" (UniqueName: \"kubernetes.io/projected/934898b1-e68f-4b10-9679-805ce7c40c02-kube-api-access-r6jmh\") on node \"crc\" DevicePath \"\"" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.748716 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.748725 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934898b1-e68f-4b10-9679-805ce7c40c02-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.938435 4838 generic.go:334] "Generic (PLEG): container finished" podID="934898b1-e68f-4b10-9679-805ce7c40c02" containerID="358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0" exitCode=0 Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.938478 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rv7l8" event={"ID":"934898b1-e68f-4b10-9679-805ce7c40c02","Type":"ContainerDied","Data":"358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0"} Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.938509 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rv7l8" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.938531 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rv7l8" event={"ID":"934898b1-e68f-4b10-9679-805ce7c40c02","Type":"ContainerDied","Data":"a279cd4d05ada4e6b43b109536066db0f824fef7a08a458c7c93e60e449a91fe"} Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.938553 4838 scope.go:117] "RemoveContainer" containerID="358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.965778 4838 scope.go:117] "RemoveContainer" containerID="68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.987351 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rv7l8"] Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.989646 4838 scope.go:117] "RemoveContainer" containerID="6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2" Dec 07 10:12:48 crc kubenswrapper[4838]: I1207 10:12:48.997696 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rv7l8"] Dec 07 10:12:49 crc kubenswrapper[4838]: I1207 10:12:49.036216 4838 scope.go:117] "RemoveContainer" containerID="358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0" Dec 07 10:12:49 crc kubenswrapper[4838]: E1207 10:12:49.036724 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0\": container with ID starting with 358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0 not found: ID does not exist" containerID="358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0" Dec 07 10:12:49 crc kubenswrapper[4838]: I1207 10:12:49.036766 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0"} err="failed to get container status \"358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0\": rpc error: code = NotFound desc = could not find container \"358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0\": container with ID starting with 358a10609fa5cd7059d309e4d35488ee96ad11aaabb8a07f016abb4e85802eb0 not found: ID does not exist" Dec 07 10:12:49 crc kubenswrapper[4838]: I1207 10:12:49.036791 4838 scope.go:117] "RemoveContainer" containerID="68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f" Dec 07 10:12:49 crc kubenswrapper[4838]: E1207 10:12:49.037216 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f\": container with ID starting with 68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f not found: ID does not exist" containerID="68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f" Dec 07 10:12:49 crc kubenswrapper[4838]: I1207 10:12:49.037247 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f"} err="failed to get container status \"68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f\": rpc error: code = NotFound desc = could not find container \"68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f\": container with ID starting with 68ee8db440ead8e257c4401c561f98c1c5abc41d3ecede7ce0c4dd45ba9cc36f not found: ID does not exist" Dec 07 10:12:49 crc kubenswrapper[4838]: I1207 10:12:49.037265 4838 scope.go:117] "RemoveContainer" containerID="6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2" Dec 07 10:12:49 crc kubenswrapper[4838]: E1207 10:12:49.037583 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2\": container with ID starting with 6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2 not found: ID does not exist" containerID="6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2" Dec 07 10:12:49 crc kubenswrapper[4838]: I1207 10:12:49.037605 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2"} err="failed to get container status \"6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2\": rpc error: code = NotFound desc = could not find container \"6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2\": container with ID starting with 6c0584ec56678c424bc5f0ece252cdd8e30472d72763f4af40bdf4c4d6536fa2 not found: ID does not exist" Dec 07 10:12:49 crc kubenswrapper[4838]: I1207 10:12:49.625492 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" path="/var/lib/kubelet/pods/934898b1-e68f-4b10-9679-805ce7c40c02/volumes" Dec 07 10:12:55 crc kubenswrapper[4838]: I1207 10:12:55.075039 4838 scope.go:117] "RemoveContainer" containerID="5f54bf662d2ab3b79c9131a3d2a81be12f47a4ca45e3b36a9f80497787e4f871" Dec 07 10:12:55 crc kubenswrapper[4838]: I1207 10:12:55.121115 4838 scope.go:117] "RemoveContainer" containerID="2486309f20fa32b2866b35fdf4f049838011c55d36a58422124a648cf2fe14f8" Dec 07 10:13:22 crc kubenswrapper[4838]: I1207 10:13:22.066846 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-gfljj"] Dec 07 10:13:22 crc kubenswrapper[4838]: I1207 10:13:22.075953 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-gfljj"] Dec 07 10:13:23 crc kubenswrapper[4838]: I1207 10:13:23.629748 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbd56671-dd88-4a32-9337-af3ee0c5c355" path="/var/lib/kubelet/pods/fbd56671-dd88-4a32-9337-af3ee0c5c355/volumes" Dec 07 10:13:54 crc kubenswrapper[4838]: I1207 10:13:54.493809 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:13:54 crc kubenswrapper[4838]: I1207 10:13:54.496474 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:13:55 crc kubenswrapper[4838]: I1207 10:13:55.277714 4838 scope.go:117] "RemoveContainer" containerID="61de118705e9018835413e44adb711b583f672aff9b1abb7f49eec9a8df63604" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.807051 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-llc4k"] Dec 07 10:14:17 crc kubenswrapper[4838]: E1207 10:14:17.807973 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="registry-server" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.807988 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="registry-server" Dec 07 10:14:17 crc kubenswrapper[4838]: E1207 10:14:17.808003 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="extract-utilities" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.808010 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="extract-utilities" Dec 07 10:14:17 crc kubenswrapper[4838]: E1207 10:14:17.808030 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="registry-server" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.808037 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="registry-server" Dec 07 10:14:17 crc kubenswrapper[4838]: E1207 10:14:17.808052 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="extract-content" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.808059 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="extract-content" Dec 07 10:14:17 crc kubenswrapper[4838]: E1207 10:14:17.808069 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="extract-content" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.808075 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="extract-content" Dec 07 10:14:17 crc kubenswrapper[4838]: E1207 10:14:17.808098 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="extract-utilities" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.808104 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="extract-utilities" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.808267 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="934898b1-e68f-4b10-9679-805ce7c40c02" containerName="registry-server" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.808284 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="92deb8e6-c5a8-44da-b1a4-e02ab7520b74" containerName="registry-server" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.809749 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.889331 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzrh7\" (UniqueName: \"kubernetes.io/projected/b5714a37-4fe0-4edb-a0c9-f99119c99245-kube-api-access-tzrh7\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.889675 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-catalog-content\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.889742 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-utilities\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.891751 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llc4k"] Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.991690 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-utilities\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.991808 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzrh7\" (UniqueName: \"kubernetes.io/projected/b5714a37-4fe0-4edb-a0c9-f99119c99245-kube-api-access-tzrh7\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.991930 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-catalog-content\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.992272 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-utilities\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:17 crc kubenswrapper[4838]: I1207 10:14:17.992309 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-catalog-content\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:18 crc kubenswrapper[4838]: I1207 10:14:18.019713 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzrh7\" (UniqueName: \"kubernetes.io/projected/b5714a37-4fe0-4edb-a0c9-f99119c99245-kube-api-access-tzrh7\") pod \"community-operators-llc4k\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:18 crc kubenswrapper[4838]: I1207 10:14:18.128249 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:18 crc kubenswrapper[4838]: I1207 10:14:18.680702 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llc4k"] Dec 07 10:14:18 crc kubenswrapper[4838]: I1207 10:14:18.841287 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llc4k" event={"ID":"b5714a37-4fe0-4edb-a0c9-f99119c99245","Type":"ContainerStarted","Data":"efa59c8b20f0743ee8773e726c08f48b89666a4ed28607f675c40a4e54fc3d4e"} Dec 07 10:14:19 crc kubenswrapper[4838]: I1207 10:14:19.857145 4838 generic.go:334] "Generic (PLEG): container finished" podID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerID="4cc647f545ca079bc9692272601826692266bdd8d4cec459e8d460ee8bead13c" exitCode=0 Dec 07 10:14:19 crc kubenswrapper[4838]: I1207 10:14:19.857246 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llc4k" event={"ID":"b5714a37-4fe0-4edb-a0c9-f99119c99245","Type":"ContainerDied","Data":"4cc647f545ca079bc9692272601826692266bdd8d4cec459e8d460ee8bead13c"} Dec 07 10:14:21 crc kubenswrapper[4838]: I1207 10:14:21.876605 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llc4k" event={"ID":"b5714a37-4fe0-4edb-a0c9-f99119c99245","Type":"ContainerStarted","Data":"4f0e6b3b8d199eb2d1ac84d1c45df3be961102631864a8c75b96164239390185"} Dec 07 10:14:22 crc kubenswrapper[4838]: I1207 10:14:22.886590 4838 generic.go:334] "Generic (PLEG): container finished" podID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerID="4f0e6b3b8d199eb2d1ac84d1c45df3be961102631864a8c75b96164239390185" exitCode=0 Dec 07 10:14:22 crc kubenswrapper[4838]: I1207 10:14:22.886874 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llc4k" event={"ID":"b5714a37-4fe0-4edb-a0c9-f99119c99245","Type":"ContainerDied","Data":"4f0e6b3b8d199eb2d1ac84d1c45df3be961102631864a8c75b96164239390185"} Dec 07 10:14:23 crc kubenswrapper[4838]: I1207 10:14:23.896467 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llc4k" event={"ID":"b5714a37-4fe0-4edb-a0c9-f99119c99245","Type":"ContainerStarted","Data":"87579c5a0ddad98256f7b1bdad8d3ccfb53b3f2a23328ee790128ad591592a2d"} Dec 07 10:14:23 crc kubenswrapper[4838]: I1207 10:14:23.920015 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-llc4k" podStartSLOduration=3.477001308 podStartE2EDuration="6.919989287s" podCreationTimestamp="2025-12-07 10:14:17 +0000 UTC" firstStartedPulling="2025-12-07 10:14:19.861195221 +0000 UTC m=+4076.568514258" lastFinishedPulling="2025-12-07 10:14:23.30418322 +0000 UTC m=+4080.011502237" observedRunningTime="2025-12-07 10:14:23.912409153 +0000 UTC m=+4080.619728190" watchObservedRunningTime="2025-12-07 10:14:23.919989287 +0000 UTC m=+4080.627308324" Dec 07 10:14:24 crc kubenswrapper[4838]: I1207 10:14:24.493692 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:14:24 crc kubenswrapper[4838]: I1207 10:14:24.494139 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:14:28 crc kubenswrapper[4838]: I1207 10:14:28.129321 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:28 crc kubenswrapper[4838]: I1207 10:14:28.131718 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:28 crc kubenswrapper[4838]: I1207 10:14:28.203028 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:29 crc kubenswrapper[4838]: I1207 10:14:29.357384 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.187343 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5mzxv"] Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.189757 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.204588 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5mzxv"] Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.299907 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-catalog-content\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.300293 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pdc8\" (UniqueName: \"kubernetes.io/projected/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-kube-api-access-9pdc8\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.300331 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-utilities\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.402328 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pdc8\" (UniqueName: \"kubernetes.io/projected/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-kube-api-access-9pdc8\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.402413 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-utilities\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.402462 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-catalog-content\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.403014 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-utilities\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.403086 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-catalog-content\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.426106 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pdc8\" (UniqueName: \"kubernetes.io/projected/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-kube-api-access-9pdc8\") pod \"certified-operators-5mzxv\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:30 crc kubenswrapper[4838]: I1207 10:14:30.568312 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:31 crc kubenswrapper[4838]: I1207 10:14:31.067499 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5mzxv"] Dec 07 10:14:31 crc kubenswrapper[4838]: I1207 10:14:31.977432 4838 generic.go:334] "Generic (PLEG): container finished" podID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerID="c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548" exitCode=0 Dec 07 10:14:31 crc kubenswrapper[4838]: I1207 10:14:31.977796 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzxv" event={"ID":"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120","Type":"ContainerDied","Data":"c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548"} Dec 07 10:14:31 crc kubenswrapper[4838]: I1207 10:14:31.977838 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzxv" event={"ID":"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120","Type":"ContainerStarted","Data":"b9f795b6a0c69a41b06955aef1dea2bc2ac72d796f284979241e6efbd93f81a3"} Dec 07 10:14:32 crc kubenswrapper[4838]: I1207 10:14:32.571680 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-llc4k"] Dec 07 10:14:32 crc kubenswrapper[4838]: I1207 10:14:32.572094 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-llc4k" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="registry-server" containerID="cri-o://87579c5a0ddad98256f7b1bdad8d3ccfb53b3f2a23328ee790128ad591592a2d" gracePeriod=2 Dec 07 10:14:32 crc kubenswrapper[4838]: I1207 10:14:32.988148 4838 generic.go:334] "Generic (PLEG): container finished" podID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerID="87579c5a0ddad98256f7b1bdad8d3ccfb53b3f2a23328ee790128ad591592a2d" exitCode=0 Dec 07 10:14:32 crc kubenswrapper[4838]: I1207 10:14:32.988282 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llc4k" event={"ID":"b5714a37-4fe0-4edb-a0c9-f99119c99245","Type":"ContainerDied","Data":"87579c5a0ddad98256f7b1bdad8d3ccfb53b3f2a23328ee790128ad591592a2d"} Dec 07 10:14:32 crc kubenswrapper[4838]: I1207 10:14:32.990860 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzxv" event={"ID":"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120","Type":"ContainerStarted","Data":"09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84"} Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.129990 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.158069 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-utilities\") pod \"b5714a37-4fe0-4edb-a0c9-f99119c99245\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.158244 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-catalog-content\") pod \"b5714a37-4fe0-4edb-a0c9-f99119c99245\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.158278 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzrh7\" (UniqueName: \"kubernetes.io/projected/b5714a37-4fe0-4edb-a0c9-f99119c99245-kube-api-access-tzrh7\") pod \"b5714a37-4fe0-4edb-a0c9-f99119c99245\" (UID: \"b5714a37-4fe0-4edb-a0c9-f99119c99245\") " Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.159707 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-utilities" (OuterVolumeSpecName: "utilities") pod "b5714a37-4fe0-4edb-a0c9-f99119c99245" (UID: "b5714a37-4fe0-4edb-a0c9-f99119c99245"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.165036 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5714a37-4fe0-4edb-a0c9-f99119c99245-kube-api-access-tzrh7" (OuterVolumeSpecName: "kube-api-access-tzrh7") pod "b5714a37-4fe0-4edb-a0c9-f99119c99245" (UID: "b5714a37-4fe0-4edb-a0c9-f99119c99245"). InnerVolumeSpecName "kube-api-access-tzrh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.207997 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5714a37-4fe0-4edb-a0c9-f99119c99245" (UID: "b5714a37-4fe0-4edb-a0c9-f99119c99245"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.261073 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.261429 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzrh7\" (UniqueName: \"kubernetes.io/projected/b5714a37-4fe0-4edb-a0c9-f99119c99245-kube-api-access-tzrh7\") on node \"crc\" DevicePath \"\"" Dec 07 10:14:33 crc kubenswrapper[4838]: I1207 10:14:33.261447 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5714a37-4fe0-4edb-a0c9-f99119c99245-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:14:34 crc kubenswrapper[4838]: I1207 10:14:34.001106 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llc4k" Dec 07 10:14:34 crc kubenswrapper[4838]: I1207 10:14:34.001560 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llc4k" event={"ID":"b5714a37-4fe0-4edb-a0c9-f99119c99245","Type":"ContainerDied","Data":"efa59c8b20f0743ee8773e726c08f48b89666a4ed28607f675c40a4e54fc3d4e"} Dec 07 10:14:34 crc kubenswrapper[4838]: I1207 10:14:34.001589 4838 scope.go:117] "RemoveContainer" containerID="87579c5a0ddad98256f7b1bdad8d3ccfb53b3f2a23328ee790128ad591592a2d" Dec 07 10:14:34 crc kubenswrapper[4838]: I1207 10:14:34.023346 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-llc4k"] Dec 07 10:14:34 crc kubenswrapper[4838]: I1207 10:14:34.023483 4838 scope.go:117] "RemoveContainer" containerID="4f0e6b3b8d199eb2d1ac84d1c45df3be961102631864a8c75b96164239390185" Dec 07 10:14:34 crc kubenswrapper[4838]: I1207 10:14:34.046059 4838 scope.go:117] "RemoveContainer" containerID="4cc647f545ca079bc9692272601826692266bdd8d4cec459e8d460ee8bead13c" Dec 07 10:14:34 crc kubenswrapper[4838]: I1207 10:14:34.046062 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-llc4k"] Dec 07 10:14:35 crc kubenswrapper[4838]: I1207 10:14:35.009597 4838 generic.go:334] "Generic (PLEG): container finished" podID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerID="09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84" exitCode=0 Dec 07 10:14:35 crc kubenswrapper[4838]: I1207 10:14:35.009643 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzxv" event={"ID":"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120","Type":"ContainerDied","Data":"09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84"} Dec 07 10:14:35 crc kubenswrapper[4838]: I1207 10:14:35.631331 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" path="/var/lib/kubelet/pods/b5714a37-4fe0-4edb-a0c9-f99119c99245/volumes" Dec 07 10:14:36 crc kubenswrapper[4838]: I1207 10:14:36.020868 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzxv" event={"ID":"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120","Type":"ContainerStarted","Data":"1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660"} Dec 07 10:14:36 crc kubenswrapper[4838]: I1207 10:14:36.075241 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5mzxv" podStartSLOduration=2.651740441 podStartE2EDuration="6.075218339s" podCreationTimestamp="2025-12-07 10:14:30 +0000 UTC" firstStartedPulling="2025-12-07 10:14:31.983126233 +0000 UTC m=+4088.690445250" lastFinishedPulling="2025-12-07 10:14:35.406604111 +0000 UTC m=+4092.113923148" observedRunningTime="2025-12-07 10:14:36.046412496 +0000 UTC m=+4092.753731533" watchObservedRunningTime="2025-12-07 10:14:36.075218339 +0000 UTC m=+4092.782537356" Dec 07 10:14:40 crc kubenswrapper[4838]: I1207 10:14:40.569375 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:40 crc kubenswrapper[4838]: I1207 10:14:40.571416 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:40 crc kubenswrapper[4838]: I1207 10:14:40.635768 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:41 crc kubenswrapper[4838]: I1207 10:14:41.155189 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:41 crc kubenswrapper[4838]: I1207 10:14:41.211109 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5mzxv"] Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.087173 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5mzxv" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="registry-server" containerID="cri-o://1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660" gracePeriod=2 Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.557091 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.582194 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-catalog-content\") pod \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.582308 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-utilities\") pod \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.582390 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pdc8\" (UniqueName: \"kubernetes.io/projected/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-kube-api-access-9pdc8\") pod \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\" (UID: \"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120\") " Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.582841 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-utilities" (OuterVolumeSpecName: "utilities") pod "f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" (UID: "f60fe6cc-f3a4-49f3-ac9e-3112dcea6120"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.583158 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.606798 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-kube-api-access-9pdc8" (OuterVolumeSpecName: "kube-api-access-9pdc8") pod "f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" (UID: "f60fe6cc-f3a4-49f3-ac9e-3112dcea6120"). InnerVolumeSpecName "kube-api-access-9pdc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.652110 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" (UID: "f60fe6cc-f3a4-49f3-ac9e-3112dcea6120"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.685585 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pdc8\" (UniqueName: \"kubernetes.io/projected/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-kube-api-access-9pdc8\") on node \"crc\" DevicePath \"\"" Dec 07 10:14:43 crc kubenswrapper[4838]: I1207 10:14:43.685763 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.106209 4838 generic.go:334] "Generic (PLEG): container finished" podID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerID="1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660" exitCode=0 Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.106252 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzxv" event={"ID":"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120","Type":"ContainerDied","Data":"1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660"} Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.106280 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzxv" event={"ID":"f60fe6cc-f3a4-49f3-ac9e-3112dcea6120","Type":"ContainerDied","Data":"b9f795b6a0c69a41b06955aef1dea2bc2ac72d796f284979241e6efbd93f81a3"} Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.106297 4838 scope.go:117] "RemoveContainer" containerID="1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.106419 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzxv" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.142715 4838 scope.go:117] "RemoveContainer" containerID="09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.149405 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5mzxv"] Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.176091 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5mzxv"] Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.193791 4838 scope.go:117] "RemoveContainer" containerID="c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.216530 4838 scope.go:117] "RemoveContainer" containerID="1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660" Dec 07 10:14:44 crc kubenswrapper[4838]: E1207 10:14:44.217037 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660\": container with ID starting with 1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660 not found: ID does not exist" containerID="1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.217085 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660"} err="failed to get container status \"1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660\": rpc error: code = NotFound desc = could not find container \"1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660\": container with ID starting with 1c26abf0e5c857461dece015406ea849f6b9a29617f544b3194d5c16918ff660 not found: ID does not exist" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.217113 4838 scope.go:117] "RemoveContainer" containerID="09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84" Dec 07 10:14:44 crc kubenswrapper[4838]: E1207 10:14:44.217469 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84\": container with ID starting with 09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84 not found: ID does not exist" containerID="09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.217534 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84"} err="failed to get container status \"09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84\": rpc error: code = NotFound desc = could not find container \"09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84\": container with ID starting with 09c5e61554a14e691e95778861fe4c2dcdbeaf22962d86b15bd046cef5066f84 not found: ID does not exist" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.217563 4838 scope.go:117] "RemoveContainer" containerID="c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548" Dec 07 10:14:44 crc kubenswrapper[4838]: E1207 10:14:44.217969 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548\": container with ID starting with c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548 not found: ID does not exist" containerID="c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548" Dec 07 10:14:44 crc kubenswrapper[4838]: I1207 10:14:44.217993 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548"} err="failed to get container status \"c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548\": rpc error: code = NotFound desc = could not find container \"c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548\": container with ID starting with c4aaab07f9daf29b11b2a48fbd75dcd0aed92c5000131da69aa37222e3602548 not found: ID does not exist" Dec 07 10:14:45 crc kubenswrapper[4838]: I1207 10:14:45.624748 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" path="/var/lib/kubelet/pods/f60fe6cc-f3a4-49f3-ac9e-3112dcea6120/volumes" Dec 07 10:14:54 crc kubenswrapper[4838]: I1207 10:14:54.493492 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:14:54 crc kubenswrapper[4838]: I1207 10:14:54.493905 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:14:54 crc kubenswrapper[4838]: I1207 10:14:54.493948 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 10:14:54 crc kubenswrapper[4838]: I1207 10:14:54.494657 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c12cece5226220a2db5c267c84bcf3318cdd31d03417d7b5d04de254f9f1a93"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 10:14:54 crc kubenswrapper[4838]: I1207 10:14:54.494705 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://8c12cece5226220a2db5c267c84bcf3318cdd31d03417d7b5d04de254f9f1a93" gracePeriod=600 Dec 07 10:14:55 crc kubenswrapper[4838]: I1207 10:14:55.237948 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="8c12cece5226220a2db5c267c84bcf3318cdd31d03417d7b5d04de254f9f1a93" exitCode=0 Dec 07 10:14:55 crc kubenswrapper[4838]: I1207 10:14:55.238544 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"8c12cece5226220a2db5c267c84bcf3318cdd31d03417d7b5d04de254f9f1a93"} Dec 07 10:14:55 crc kubenswrapper[4838]: I1207 10:14:55.238573 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9"} Dec 07 10:14:55 crc kubenswrapper[4838]: I1207 10:14:55.238590 4838 scope.go:117] "RemoveContainer" containerID="d184c87df0cb0b5cde63971f3a49e57b1ab0c9b8535f1f834613ac80c1dc1ef3" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.173718 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb"] Dec 07 10:15:00 crc kubenswrapper[4838]: E1207 10:15:00.175842 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="extract-utilities" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.175935 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="extract-utilities" Dec 07 10:15:00 crc kubenswrapper[4838]: E1207 10:15:00.176043 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="extract-content" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.176141 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="extract-content" Dec 07 10:15:00 crc kubenswrapper[4838]: E1207 10:15:00.176231 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="extract-content" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.176300 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="extract-content" Dec 07 10:15:00 crc kubenswrapper[4838]: E1207 10:15:00.176381 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="extract-utilities" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.176450 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="extract-utilities" Dec 07 10:15:00 crc kubenswrapper[4838]: E1207 10:15:00.176516 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="registry-server" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.176583 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="registry-server" Dec 07 10:15:00 crc kubenswrapper[4838]: E1207 10:15:00.176652 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="registry-server" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.176716 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="registry-server" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.177024 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5714a37-4fe0-4edb-a0c9-f99119c99245" containerName="registry-server" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.177131 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f60fe6cc-f3a4-49f3-ac9e-3112dcea6120" containerName="registry-server" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.177955 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.180507 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.180745 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.184353 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb"] Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.321944 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/88243bac-65e8-4100-bbec-f4cf0a842bc0-secret-volume\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.322060 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88243bac-65e8-4100-bbec-f4cf0a842bc0-config-volume\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.322432 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd949\" (UniqueName: \"kubernetes.io/projected/88243bac-65e8-4100-bbec-f4cf0a842bc0-kube-api-access-dd949\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.424483 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88243bac-65e8-4100-bbec-f4cf0a842bc0-config-volume\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.424631 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd949\" (UniqueName: \"kubernetes.io/projected/88243bac-65e8-4100-bbec-f4cf0a842bc0-kube-api-access-dd949\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.424682 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/88243bac-65e8-4100-bbec-f4cf0a842bc0-secret-volume\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.425727 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88243bac-65e8-4100-bbec-f4cf0a842bc0-config-volume\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.434606 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/88243bac-65e8-4100-bbec-f4cf0a842bc0-secret-volume\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.442885 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd949\" (UniqueName: \"kubernetes.io/projected/88243bac-65e8-4100-bbec-f4cf0a842bc0-kube-api-access-dd949\") pod \"collect-profiles-29418375-gfsbb\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:00 crc kubenswrapper[4838]: I1207 10:15:00.530120 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:01 crc kubenswrapper[4838]: I1207 10:15:01.039376 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb"] Dec 07 10:15:01 crc kubenswrapper[4838]: I1207 10:15:01.307138 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" event={"ID":"88243bac-65e8-4100-bbec-f4cf0a842bc0","Type":"ContainerStarted","Data":"7cc95816b6e73124fa7d6ade64b45aca9d86eb7235169f52409ad436e4b76f4b"} Dec 07 10:15:02 crc kubenswrapper[4838]: I1207 10:15:02.319885 4838 generic.go:334] "Generic (PLEG): container finished" podID="88243bac-65e8-4100-bbec-f4cf0a842bc0" containerID="630c702b8d6113693cd39d901e47d6c69d773b34f9f075951f49582fbef6f924" exitCode=0 Dec 07 10:15:02 crc kubenswrapper[4838]: I1207 10:15:02.319995 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" event={"ID":"88243bac-65e8-4100-bbec-f4cf0a842bc0","Type":"ContainerDied","Data":"630c702b8d6113693cd39d901e47d6c69d773b34f9f075951f49582fbef6f924"} Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.770143 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.899484 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88243bac-65e8-4100-bbec-f4cf0a842bc0-config-volume\") pod \"88243bac-65e8-4100-bbec-f4cf0a842bc0\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.899931 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/88243bac-65e8-4100-bbec-f4cf0a842bc0-secret-volume\") pod \"88243bac-65e8-4100-bbec-f4cf0a842bc0\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.900942 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88243bac-65e8-4100-bbec-f4cf0a842bc0-config-volume" (OuterVolumeSpecName: "config-volume") pod "88243bac-65e8-4100-bbec-f4cf0a842bc0" (UID: "88243bac-65e8-4100-bbec-f4cf0a842bc0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.901098 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd949\" (UniqueName: \"kubernetes.io/projected/88243bac-65e8-4100-bbec-f4cf0a842bc0-kube-api-access-dd949\") pod \"88243bac-65e8-4100-bbec-f4cf0a842bc0\" (UID: \"88243bac-65e8-4100-bbec-f4cf0a842bc0\") " Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.901602 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88243bac-65e8-4100-bbec-f4cf0a842bc0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.905784 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88243bac-65e8-4100-bbec-f4cf0a842bc0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "88243bac-65e8-4100-bbec-f4cf0a842bc0" (UID: "88243bac-65e8-4100-bbec-f4cf0a842bc0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:15:03 crc kubenswrapper[4838]: I1207 10:15:03.906376 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88243bac-65e8-4100-bbec-f4cf0a842bc0-kube-api-access-dd949" (OuterVolumeSpecName: "kube-api-access-dd949") pod "88243bac-65e8-4100-bbec-f4cf0a842bc0" (UID: "88243bac-65e8-4100-bbec-f4cf0a842bc0"). InnerVolumeSpecName "kube-api-access-dd949". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:15:04 crc kubenswrapper[4838]: I1207 10:15:04.004505 4838 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/88243bac-65e8-4100-bbec-f4cf0a842bc0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 10:15:04 crc kubenswrapper[4838]: I1207 10:15:04.004544 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd949\" (UniqueName: \"kubernetes.io/projected/88243bac-65e8-4100-bbec-f4cf0a842bc0-kube-api-access-dd949\") on node \"crc\" DevicePath \"\"" Dec 07 10:15:04 crc kubenswrapper[4838]: I1207 10:15:04.339039 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" event={"ID":"88243bac-65e8-4100-bbec-f4cf0a842bc0","Type":"ContainerDied","Data":"7cc95816b6e73124fa7d6ade64b45aca9d86eb7235169f52409ad436e4b76f4b"} Dec 07 10:15:04 crc kubenswrapper[4838]: I1207 10:15:04.339562 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418375-gfsbb" Dec 07 10:15:04 crc kubenswrapper[4838]: I1207 10:15:04.339363 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cc95816b6e73124fa7d6ade64b45aca9d86eb7235169f52409ad436e4b76f4b" Dec 07 10:15:04 crc kubenswrapper[4838]: I1207 10:15:04.865711 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t"] Dec 07 10:15:04 crc kubenswrapper[4838]: I1207 10:15:04.875872 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418330-gbw9t"] Dec 07 10:15:05 crc kubenswrapper[4838]: I1207 10:15:05.629457 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e88c0c3-c7da-469e-b34e-7e2d7672611c" path="/var/lib/kubelet/pods/7e88c0c3-c7da-469e-b34e-7e2d7672611c/volumes" Dec 07 10:15:55 crc kubenswrapper[4838]: I1207 10:15:55.380523 4838 scope.go:117] "RemoveContainer" containerID="a8d433a11c4db30548a4e54a6a8e6119f2415d8689c13931367e21a2c50e97bd" Dec 07 10:16:54 crc kubenswrapper[4838]: I1207 10:16:54.493611 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:16:54 crc kubenswrapper[4838]: I1207 10:16:54.494170 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:17:24 crc kubenswrapper[4838]: I1207 10:17:24.493144 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:17:24 crc kubenswrapper[4838]: I1207 10:17:24.493601 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:17:45 crc kubenswrapper[4838]: I1207 10:17:45.007409 4838 generic.go:334] "Generic (PLEG): container finished" podID="a430a791-ad62-4246-8422-6bae0b6ce772" containerID="73700bfabf6d58d2ddafd40f5b882a22f9a570bb72e32c37029a1a76af6b9805" exitCode=0 Dec 07 10:17:45 crc kubenswrapper[4838]: I1207 10:17:45.007536 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a430a791-ad62-4246-8422-6bae0b6ce772","Type":"ContainerDied","Data":"73700bfabf6d58d2ddafd40f5b882a22f9a570bb72e32c37029a1a76af6b9805"} Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.387016 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528035 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config-secret\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528087 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ssh-key\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528227 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-config-data\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528256 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtb2k\" (UniqueName: \"kubernetes.io/projected/a430a791-ad62-4246-8422-6bae0b6ce772-kube-api-access-mtb2k\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528280 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ca-certs\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528319 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528346 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528454 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-workdir\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.528554 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-temporary\") pod \"a430a791-ad62-4246-8422-6bae0b6ce772\" (UID: \"a430a791-ad62-4246-8422-6bae0b6ce772\") " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.529569 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-config-data" (OuterVolumeSpecName: "config-data") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.529687 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.530102 4838 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.530118 4838 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.534197 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.535064 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a430a791-ad62-4246-8422-6bae0b6ce772-kube-api-access-mtb2k" (OuterVolumeSpecName: "kube-api-access-mtb2k") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "kube-api-access-mtb2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.545420 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.557511 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.561256 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.581023 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.583425 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a430a791-ad62-4246-8422-6bae0b6ce772" (UID: "a430a791-ad62-4246-8422-6bae0b6ce772"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.634902 4838 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a430a791-ad62-4246-8422-6bae0b6ce772-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.635068 4838 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.635163 4838 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.635228 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtb2k\" (UniqueName: \"kubernetes.io/projected/a430a791-ad62-4246-8422-6bae0b6ce772-kube-api-access-mtb2k\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.635287 4838 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a430a791-ad62-4246-8422-6bae0b6ce772-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.635344 4838 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a430a791-ad62-4246-8422-6bae0b6ce772-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.635421 4838 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.658242 4838 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 07 10:17:46 crc kubenswrapper[4838]: I1207 10:17:46.737431 4838 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 07 10:17:47 crc kubenswrapper[4838]: I1207 10:17:47.040622 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a430a791-ad62-4246-8422-6bae0b6ce772","Type":"ContainerDied","Data":"8f3b31953ba8c0eb4c530a1df4c35a35649b0005e492511bec032ae8eac4acbc"} Dec 07 10:17:47 crc kubenswrapper[4838]: I1207 10:17:47.040881 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f3b31953ba8c0eb4c530a1df4c35a35649b0005e492511bec032ae8eac4acbc" Dec 07 10:17:47 crc kubenswrapper[4838]: I1207 10:17:47.040735 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 10:17:47 crc kubenswrapper[4838]: E1207 10:17:47.229463 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda430a791_ad62_4246_8422_6bae0b6ce772.slice\": RecentStats: unable to find data in memory cache]" Dec 07 10:17:54 crc kubenswrapper[4838]: I1207 10:17:54.493423 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:17:54 crc kubenswrapper[4838]: I1207 10:17:54.493923 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:17:54 crc kubenswrapper[4838]: I1207 10:17:54.493974 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 10:17:54 crc kubenswrapper[4838]: I1207 10:17:54.494796 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 10:17:54 crc kubenswrapper[4838]: I1207 10:17:54.494883 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" gracePeriod=600 Dec 07 10:17:54 crc kubenswrapper[4838]: E1207 10:17:54.621004 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:17:55 crc kubenswrapper[4838]: I1207 10:17:55.124327 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" exitCode=0 Dec 07 10:17:55 crc kubenswrapper[4838]: I1207 10:17:55.124436 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9"} Dec 07 10:17:55 crc kubenswrapper[4838]: I1207 10:17:55.124651 4838 scope.go:117] "RemoveContainer" containerID="8c12cece5226220a2db5c267c84bcf3318cdd31d03417d7b5d04de254f9f1a93" Dec 07 10:17:55 crc kubenswrapper[4838]: I1207 10:17:55.125607 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:17:55 crc kubenswrapper[4838]: E1207 10:17:55.126066 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.632151 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 07 10:17:58 crc kubenswrapper[4838]: E1207 10:17:58.632729 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a430a791-ad62-4246-8422-6bae0b6ce772" containerName="tempest-tests-tempest-tests-runner" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.632743 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="a430a791-ad62-4246-8422-6bae0b6ce772" containerName="tempest-tests-tempest-tests-runner" Dec 07 10:17:58 crc kubenswrapper[4838]: E1207 10:17:58.632770 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88243bac-65e8-4100-bbec-f4cf0a842bc0" containerName="collect-profiles" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.632777 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="88243bac-65e8-4100-bbec-f4cf0a842bc0" containerName="collect-profiles" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.632974 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="a430a791-ad62-4246-8422-6bae0b6ce772" containerName="tempest-tests-tempest-tests-runner" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.632998 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="88243bac-65e8-4100-bbec-f4cf0a842bc0" containerName="collect-profiles" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.633540 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.636042 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pcr24" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.645892 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.781940 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m885d\" (UniqueName: \"kubernetes.io/projected/17d4b7b1-e69f-444b-bd80-753b07d66af7-kube-api-access-m885d\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"17d4b7b1-e69f-444b-bd80-753b07d66af7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.782014 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"17d4b7b1-e69f-444b-bd80-753b07d66af7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.883679 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m885d\" (UniqueName: \"kubernetes.io/projected/17d4b7b1-e69f-444b-bd80-753b07d66af7-kube-api-access-m885d\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"17d4b7b1-e69f-444b-bd80-753b07d66af7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.884222 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"17d4b7b1-e69f-444b-bd80-753b07d66af7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.884712 4838 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"17d4b7b1-e69f-444b-bd80-753b07d66af7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.908296 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m885d\" (UniqueName: \"kubernetes.io/projected/17d4b7b1-e69f-444b-bd80-753b07d66af7-kube-api-access-m885d\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"17d4b7b1-e69f-444b-bd80-753b07d66af7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:58 crc kubenswrapper[4838]: I1207 10:17:58.912872 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"17d4b7b1-e69f-444b-bd80-753b07d66af7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:59 crc kubenswrapper[4838]: I1207 10:17:59.020947 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 10:17:59 crc kubenswrapper[4838]: I1207 10:17:59.467106 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 07 10:17:59 crc kubenswrapper[4838]: I1207 10:17:59.483015 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 10:18:00 crc kubenswrapper[4838]: I1207 10:18:00.182779 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"17d4b7b1-e69f-444b-bd80-753b07d66af7","Type":"ContainerStarted","Data":"7de25e78c6d42b9b6e726ddea93a1ea4c866552c6639d95d4f708ead64e7f406"} Dec 07 10:18:01 crc kubenswrapper[4838]: I1207 10:18:01.192247 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"17d4b7b1-e69f-444b-bd80-753b07d66af7","Type":"ContainerStarted","Data":"31c46604867e1f0a1c4596525560869461312123f8ca5898e9f06fb21549cc73"} Dec 07 10:18:01 crc kubenswrapper[4838]: I1207 10:18:01.212465 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.9952743499999999 podStartE2EDuration="3.212439887s" podCreationTimestamp="2025-12-07 10:17:58 +0000 UTC" firstStartedPulling="2025-12-07 10:17:59.482751137 +0000 UTC m=+4296.190070164" lastFinishedPulling="2025-12-07 10:18:00.699916684 +0000 UTC m=+4297.407235701" observedRunningTime="2025-12-07 10:18:01.206612492 +0000 UTC m=+4297.913931519" watchObservedRunningTime="2025-12-07 10:18:01.212439887 +0000 UTC m=+4297.919758924" Dec 07 10:18:05 crc kubenswrapper[4838]: I1207 10:18:05.614401 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:18:05 crc kubenswrapper[4838]: E1207 10:18:05.615265 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:18:19 crc kubenswrapper[4838]: I1207 10:18:19.615994 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:18:19 crc kubenswrapper[4838]: E1207 10:18:19.616782 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.175937 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tjr69/must-gather-sqwk6"] Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.177774 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.179787 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tjr69"/"default-dockercfg-bqqjl" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.183724 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tjr69"/"kube-root-ca.crt" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.186292 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tjr69"/"openshift-service-ca.crt" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.196139 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tjr69/must-gather-sqwk6"] Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.302064 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa0a6d32-e204-43ce-b6a0-609b93bea035-must-gather-output\") pod \"must-gather-sqwk6\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.302152 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv447\" (UniqueName: \"kubernetes.io/projected/fa0a6d32-e204-43ce-b6a0-609b93bea035-kube-api-access-cv447\") pod \"must-gather-sqwk6\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.403636 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa0a6d32-e204-43ce-b6a0-609b93bea035-must-gather-output\") pod \"must-gather-sqwk6\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.403701 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv447\" (UniqueName: \"kubernetes.io/projected/fa0a6d32-e204-43ce-b6a0-609b93bea035-kube-api-access-cv447\") pod \"must-gather-sqwk6\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.404214 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa0a6d32-e204-43ce-b6a0-609b93bea035-must-gather-output\") pod \"must-gather-sqwk6\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.420687 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv447\" (UniqueName: \"kubernetes.io/projected/fa0a6d32-e204-43ce-b6a0-609b93bea035-kube-api-access-cv447\") pod \"must-gather-sqwk6\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.493524 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:18:25 crc kubenswrapper[4838]: I1207 10:18:25.960460 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tjr69/must-gather-sqwk6"] Dec 07 10:18:26 crc kubenswrapper[4838]: W1207 10:18:26.418287 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa0a6d32_e204_43ce_b6a0_609b93bea035.slice/crio-acaaf685b68099d59a46ebdbd175d5ea5f277b77e7e7c55d5c1c0c1a172d36a9 WatchSource:0}: Error finding container acaaf685b68099d59a46ebdbd175d5ea5f277b77e7e7c55d5c1c0c1a172d36a9: Status 404 returned error can't find the container with id acaaf685b68099d59a46ebdbd175d5ea5f277b77e7e7c55d5c1c0c1a172d36a9 Dec 07 10:18:26 crc kubenswrapper[4838]: I1207 10:18:26.521518 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/must-gather-sqwk6" event={"ID":"fa0a6d32-e204-43ce-b6a0-609b93bea035","Type":"ContainerStarted","Data":"acaaf685b68099d59a46ebdbd175d5ea5f277b77e7e7c55d5c1c0c1a172d36a9"} Dec 07 10:18:31 crc kubenswrapper[4838]: I1207 10:18:31.579981 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/must-gather-sqwk6" event={"ID":"fa0a6d32-e204-43ce-b6a0-609b93bea035","Type":"ContainerStarted","Data":"7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c"} Dec 07 10:18:31 crc kubenswrapper[4838]: I1207 10:18:31.580482 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/must-gather-sqwk6" event={"ID":"fa0a6d32-e204-43ce-b6a0-609b93bea035","Type":"ContainerStarted","Data":"46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633"} Dec 07 10:18:31 crc kubenswrapper[4838]: I1207 10:18:31.607030 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tjr69/must-gather-sqwk6" podStartSLOduration=2.0913874 podStartE2EDuration="6.607009317s" podCreationTimestamp="2025-12-07 10:18:25 +0000 UTC" firstStartedPulling="2025-12-07 10:18:26.421744733 +0000 UTC m=+4323.129063760" lastFinishedPulling="2025-12-07 10:18:30.93736666 +0000 UTC m=+4327.644685677" observedRunningTime="2025-12-07 10:18:31.597554271 +0000 UTC m=+4328.304873298" watchObservedRunningTime="2025-12-07 10:18:31.607009317 +0000 UTC m=+4328.314328344" Dec 07 10:18:32 crc kubenswrapper[4838]: I1207 10:18:32.615871 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:18:32 crc kubenswrapper[4838]: E1207 10:18:32.616538 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.628507 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tjr69/crc-debug-w8czc"] Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.633646 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.764936 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c4254-21ca-4f54-86cd-d1a950ac3a51-host\") pod \"crc-debug-w8czc\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.765118 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bxn4\" (UniqueName: \"kubernetes.io/projected/208c4254-21ca-4f54-86cd-d1a950ac3a51-kube-api-access-5bxn4\") pod \"crc-debug-w8czc\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.867218 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bxn4\" (UniqueName: \"kubernetes.io/projected/208c4254-21ca-4f54-86cd-d1a950ac3a51-kube-api-access-5bxn4\") pod \"crc-debug-w8czc\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.867352 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c4254-21ca-4f54-86cd-d1a950ac3a51-host\") pod \"crc-debug-w8czc\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.867487 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c4254-21ca-4f54-86cd-d1a950ac3a51-host\") pod \"crc-debug-w8czc\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.886962 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bxn4\" (UniqueName: \"kubernetes.io/projected/208c4254-21ca-4f54-86cd-d1a950ac3a51-kube-api-access-5bxn4\") pod \"crc-debug-w8czc\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:36 crc kubenswrapper[4838]: I1207 10:18:36.962571 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:18:37 crc kubenswrapper[4838]: I1207 10:18:37.655842 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-w8czc" event={"ID":"208c4254-21ca-4f54-86cd-d1a950ac3a51","Type":"ContainerStarted","Data":"8220bba66c6c3daadf27e303fd728d798a33a3a5cb6d8a32fb1927d1f3ce127f"} Dec 07 10:18:45 crc kubenswrapper[4838]: I1207 10:18:45.634318 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:18:45 crc kubenswrapper[4838]: E1207 10:18:45.634953 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:18:48 crc kubenswrapper[4838]: I1207 10:18:48.767450 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-w8czc" event={"ID":"208c4254-21ca-4f54-86cd-d1a950ac3a51","Type":"ContainerStarted","Data":"8b10203985b24b0953946ac2af1394dba4365c8a756f678cdbf635d3ca9120e4"} Dec 07 10:18:48 crc kubenswrapper[4838]: I1207 10:18:48.794873 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tjr69/crc-debug-w8czc" podStartSLOduration=2.05304021 podStartE2EDuration="12.794853774s" podCreationTimestamp="2025-12-07 10:18:36 +0000 UTC" firstStartedPulling="2025-12-07 10:18:37.299488935 +0000 UTC m=+4334.006807952" lastFinishedPulling="2025-12-07 10:18:48.041302499 +0000 UTC m=+4344.748621516" observedRunningTime="2025-12-07 10:18:48.783395111 +0000 UTC m=+4345.490714138" watchObservedRunningTime="2025-12-07 10:18:48.794853774 +0000 UTC m=+4345.502172791" Dec 07 10:19:00 crc kubenswrapper[4838]: I1207 10:19:00.615633 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:19:00 crc kubenswrapper[4838]: E1207 10:19:00.616381 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:19:13 crc kubenswrapper[4838]: I1207 10:19:13.621429 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:19:13 crc kubenswrapper[4838]: E1207 10:19:13.623237 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:19:25 crc kubenswrapper[4838]: I1207 10:19:25.614883 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:19:25 crc kubenswrapper[4838]: E1207 10:19:25.616331 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:19:29 crc kubenswrapper[4838]: E1207 10:19:29.969666 4838 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod208c4254_21ca_4f54_86cd_d1a950ac3a51.slice/crio-conmon-8b10203985b24b0953946ac2af1394dba4365c8a756f678cdbf635d3ca9120e4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod208c4254_21ca_4f54_86cd_d1a950ac3a51.slice/crio-8b10203985b24b0953946ac2af1394dba4365c8a756f678cdbf635d3ca9120e4.scope\": RecentStats: unable to find data in memory cache]" Dec 07 10:19:30 crc kubenswrapper[4838]: I1207 10:19:30.140585 4838 generic.go:334] "Generic (PLEG): container finished" podID="208c4254-21ca-4f54-86cd-d1a950ac3a51" containerID="8b10203985b24b0953946ac2af1394dba4365c8a756f678cdbf635d3ca9120e4" exitCode=0 Dec 07 10:19:30 crc kubenswrapper[4838]: I1207 10:19:30.140991 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-w8czc" event={"ID":"208c4254-21ca-4f54-86cd-d1a950ac3a51","Type":"ContainerDied","Data":"8b10203985b24b0953946ac2af1394dba4365c8a756f678cdbf635d3ca9120e4"} Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.265484 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.307959 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tjr69/crc-debug-w8czc"] Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.323421 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tjr69/crc-debug-w8czc"] Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.419399 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c4254-21ca-4f54-86cd-d1a950ac3a51-host\") pod \"208c4254-21ca-4f54-86cd-d1a950ac3a51\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.419480 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bxn4\" (UniqueName: \"kubernetes.io/projected/208c4254-21ca-4f54-86cd-d1a950ac3a51-kube-api-access-5bxn4\") pod \"208c4254-21ca-4f54-86cd-d1a950ac3a51\" (UID: \"208c4254-21ca-4f54-86cd-d1a950ac3a51\") " Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.420005 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/208c4254-21ca-4f54-86cd-d1a950ac3a51-host" (OuterVolumeSpecName: "host") pod "208c4254-21ca-4f54-86cd-d1a950ac3a51" (UID: "208c4254-21ca-4f54-86cd-d1a950ac3a51"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.420337 4838 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/208c4254-21ca-4f54-86cd-d1a950ac3a51-host\") on node \"crc\" DevicePath \"\"" Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.431975 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/208c4254-21ca-4f54-86cd-d1a950ac3a51-kube-api-access-5bxn4" (OuterVolumeSpecName: "kube-api-access-5bxn4") pod "208c4254-21ca-4f54-86cd-d1a950ac3a51" (UID: "208c4254-21ca-4f54-86cd-d1a950ac3a51"). InnerVolumeSpecName "kube-api-access-5bxn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.522257 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bxn4\" (UniqueName: \"kubernetes.io/projected/208c4254-21ca-4f54-86cd-d1a950ac3a51-kube-api-access-5bxn4\") on node \"crc\" DevicePath \"\"" Dec 07 10:19:31 crc kubenswrapper[4838]: I1207 10:19:31.646573 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="208c4254-21ca-4f54-86cd-d1a950ac3a51" path="/var/lib/kubelet/pods/208c4254-21ca-4f54-86cd-d1a950ac3a51/volumes" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.158355 4838 scope.go:117] "RemoveContainer" containerID="8b10203985b24b0953946ac2af1394dba4365c8a756f678cdbf635d3ca9120e4" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.158409 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-w8czc" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.537566 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tjr69/crc-debug-sfg4f"] Dec 07 10:19:32 crc kubenswrapper[4838]: E1207 10:19:32.538250 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208c4254-21ca-4f54-86cd-d1a950ac3a51" containerName="container-00" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.538263 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="208c4254-21ca-4f54-86cd-d1a950ac3a51" containerName="container-00" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.538458 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="208c4254-21ca-4f54-86cd-d1a950ac3a51" containerName="container-00" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.539077 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.642314 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6577n\" (UniqueName: \"kubernetes.io/projected/3ce45682-6c0d-44a9-b01f-d48736427c89-kube-api-access-6577n\") pod \"crc-debug-sfg4f\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.642550 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ce45682-6c0d-44a9-b01f-d48736427c89-host\") pod \"crc-debug-sfg4f\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.744773 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ce45682-6c0d-44a9-b01f-d48736427c89-host\") pod \"crc-debug-sfg4f\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.744907 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6577n\" (UniqueName: \"kubernetes.io/projected/3ce45682-6c0d-44a9-b01f-d48736427c89-kube-api-access-6577n\") pod \"crc-debug-sfg4f\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.744930 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ce45682-6c0d-44a9-b01f-d48736427c89-host\") pod \"crc-debug-sfg4f\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.788851 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6577n\" (UniqueName: \"kubernetes.io/projected/3ce45682-6c0d-44a9-b01f-d48736427c89-kube-api-access-6577n\") pod \"crc-debug-sfg4f\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:32 crc kubenswrapper[4838]: I1207 10:19:32.855631 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:33 crc kubenswrapper[4838]: I1207 10:19:33.167499 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" event={"ID":"3ce45682-6c0d-44a9-b01f-d48736427c89","Type":"ContainerStarted","Data":"ec2588ab369494e1ccdb2d36f5886154eccc05d388ac44c4a5b1aa1a4c2845a9"} Dec 07 10:19:33 crc kubenswrapper[4838]: I1207 10:19:33.167839 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" event={"ID":"3ce45682-6c0d-44a9-b01f-d48736427c89","Type":"ContainerStarted","Data":"955bc653838b3c160b46abd04bc5100389454d7449dc9ab1183776c4c0fa3d35"} Dec 07 10:19:33 crc kubenswrapper[4838]: I1207 10:19:33.182492 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" podStartSLOduration=1.182473338 podStartE2EDuration="1.182473338s" podCreationTimestamp="2025-12-07 10:19:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:19:33.179655339 +0000 UTC m=+4389.886974366" watchObservedRunningTime="2025-12-07 10:19:33.182473338 +0000 UTC m=+4389.889792355" Dec 07 10:19:34 crc kubenswrapper[4838]: I1207 10:19:34.177310 4838 generic.go:334] "Generic (PLEG): container finished" podID="3ce45682-6c0d-44a9-b01f-d48736427c89" containerID="ec2588ab369494e1ccdb2d36f5886154eccc05d388ac44c4a5b1aa1a4c2845a9" exitCode=0 Dec 07 10:19:34 crc kubenswrapper[4838]: I1207 10:19:34.177406 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" event={"ID":"3ce45682-6c0d-44a9-b01f-d48736427c89","Type":"ContainerDied","Data":"ec2588ab369494e1ccdb2d36f5886154eccc05d388ac44c4a5b1aa1a4c2845a9"} Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.395993 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.423713 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tjr69/crc-debug-sfg4f"] Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.431303 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tjr69/crc-debug-sfg4f"] Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.495076 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6577n\" (UniqueName: \"kubernetes.io/projected/3ce45682-6c0d-44a9-b01f-d48736427c89-kube-api-access-6577n\") pod \"3ce45682-6c0d-44a9-b01f-d48736427c89\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.495276 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ce45682-6c0d-44a9-b01f-d48736427c89-host\") pod \"3ce45682-6c0d-44a9-b01f-d48736427c89\" (UID: \"3ce45682-6c0d-44a9-b01f-d48736427c89\") " Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.495346 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ce45682-6c0d-44a9-b01f-d48736427c89-host" (OuterVolumeSpecName: "host") pod "3ce45682-6c0d-44a9-b01f-d48736427c89" (UID: "3ce45682-6c0d-44a9-b01f-d48736427c89"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.495675 4838 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ce45682-6c0d-44a9-b01f-d48736427c89-host\") on node \"crc\" DevicePath \"\"" Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.507983 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ce45682-6c0d-44a9-b01f-d48736427c89-kube-api-access-6577n" (OuterVolumeSpecName: "kube-api-access-6577n") pod "3ce45682-6c0d-44a9-b01f-d48736427c89" (UID: "3ce45682-6c0d-44a9-b01f-d48736427c89"). InnerVolumeSpecName "kube-api-access-6577n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.597184 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6577n\" (UniqueName: \"kubernetes.io/projected/3ce45682-6c0d-44a9-b01f-d48736427c89-kube-api-access-6577n\") on node \"crc\" DevicePath \"\"" Dec 07 10:19:35 crc kubenswrapper[4838]: I1207 10:19:35.624944 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ce45682-6c0d-44a9-b01f-d48736427c89" path="/var/lib/kubelet/pods/3ce45682-6c0d-44a9-b01f-d48736427c89/volumes" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.197106 4838 scope.go:117] "RemoveContainer" containerID="ec2588ab369494e1ccdb2d36f5886154eccc05d388ac44c4a5b1aa1a4c2845a9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.197445 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-sfg4f" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.646348 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tjr69/crc-debug-mq8k9"] Dec 07 10:19:36 crc kubenswrapper[4838]: E1207 10:19:36.648005 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ce45682-6c0d-44a9-b01f-d48736427c89" containerName="container-00" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.648123 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ce45682-6c0d-44a9-b01f-d48736427c89" containerName="container-00" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.648465 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ce45682-6c0d-44a9-b01f-d48736427c89" containerName="container-00" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.649326 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.721084 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4zkp\" (UniqueName: \"kubernetes.io/projected/6263c5b5-1a1e-4840-b3a1-2241162c05d9-kube-api-access-g4zkp\") pod \"crc-debug-mq8k9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.721194 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6263c5b5-1a1e-4840-b3a1-2241162c05d9-host\") pod \"crc-debug-mq8k9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.822957 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4zkp\" (UniqueName: \"kubernetes.io/projected/6263c5b5-1a1e-4840-b3a1-2241162c05d9-kube-api-access-g4zkp\") pod \"crc-debug-mq8k9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.823017 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6263c5b5-1a1e-4840-b3a1-2241162c05d9-host\") pod \"crc-debug-mq8k9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.823127 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6263c5b5-1a1e-4840-b3a1-2241162c05d9-host\") pod \"crc-debug-mq8k9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.842257 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4zkp\" (UniqueName: \"kubernetes.io/projected/6263c5b5-1a1e-4840-b3a1-2241162c05d9-kube-api-access-g4zkp\") pod \"crc-debug-mq8k9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:36 crc kubenswrapper[4838]: I1207 10:19:36.977250 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:37 crc kubenswrapper[4838]: I1207 10:19:37.208861 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-mq8k9" event={"ID":"6263c5b5-1a1e-4840-b3a1-2241162c05d9","Type":"ContainerStarted","Data":"fe556c0cf202640ff5fe7ed6e0ea5d7a37c53eeaa2ca4b3f76079c6e6b5cc39b"} Dec 07 10:19:38 crc kubenswrapper[4838]: I1207 10:19:38.217293 4838 generic.go:334] "Generic (PLEG): container finished" podID="6263c5b5-1a1e-4840-b3a1-2241162c05d9" containerID="170125501ff9314cf1e126a93254e32c2f5bf85f21e3142474f876b903f82c43" exitCode=0 Dec 07 10:19:38 crc kubenswrapper[4838]: I1207 10:19:38.217334 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/crc-debug-mq8k9" event={"ID":"6263c5b5-1a1e-4840-b3a1-2241162c05d9","Type":"ContainerDied","Data":"170125501ff9314cf1e126a93254e32c2f5bf85f21e3142474f876b903f82c43"} Dec 07 10:19:38 crc kubenswrapper[4838]: I1207 10:19:38.259680 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tjr69/crc-debug-mq8k9"] Dec 07 10:19:38 crc kubenswrapper[4838]: I1207 10:19:38.269368 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tjr69/crc-debug-mq8k9"] Dec 07 10:19:38 crc kubenswrapper[4838]: I1207 10:19:38.614558 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:19:38 crc kubenswrapper[4838]: E1207 10:19:38.615339 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.331430 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.370960 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4zkp\" (UniqueName: \"kubernetes.io/projected/6263c5b5-1a1e-4840-b3a1-2241162c05d9-kube-api-access-g4zkp\") pod \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.371049 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6263c5b5-1a1e-4840-b3a1-2241162c05d9-host\") pod \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\" (UID: \"6263c5b5-1a1e-4840-b3a1-2241162c05d9\") " Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.371160 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6263c5b5-1a1e-4840-b3a1-2241162c05d9-host" (OuterVolumeSpecName: "host") pod "6263c5b5-1a1e-4840-b3a1-2241162c05d9" (UID: "6263c5b5-1a1e-4840-b3a1-2241162c05d9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.371370 4838 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6263c5b5-1a1e-4840-b3a1-2241162c05d9-host\") on node \"crc\" DevicePath \"\"" Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.376835 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6263c5b5-1a1e-4840-b3a1-2241162c05d9-kube-api-access-g4zkp" (OuterVolumeSpecName: "kube-api-access-g4zkp") pod "6263c5b5-1a1e-4840-b3a1-2241162c05d9" (UID: "6263c5b5-1a1e-4840-b3a1-2241162c05d9"). InnerVolumeSpecName "kube-api-access-g4zkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.472751 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4zkp\" (UniqueName: \"kubernetes.io/projected/6263c5b5-1a1e-4840-b3a1-2241162c05d9-kube-api-access-g4zkp\") on node \"crc\" DevicePath \"\"" Dec 07 10:19:39 crc kubenswrapper[4838]: I1207 10:19:39.629993 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6263c5b5-1a1e-4840-b3a1-2241162c05d9" path="/var/lib/kubelet/pods/6263c5b5-1a1e-4840-b3a1-2241162c05d9/volumes" Dec 07 10:19:40 crc kubenswrapper[4838]: I1207 10:19:40.274367 4838 scope.go:117] "RemoveContainer" containerID="170125501ff9314cf1e126a93254e32c2f5bf85f21e3142474f876b903f82c43" Dec 07 10:19:40 crc kubenswrapper[4838]: I1207 10:19:40.274510 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/crc-debug-mq8k9" Dec 07 10:19:53 crc kubenswrapper[4838]: I1207 10:19:53.620726 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:19:53 crc kubenswrapper[4838]: E1207 10:19:53.621530 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:20:08 crc kubenswrapper[4838]: I1207 10:20:08.614236 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:20:08 crc kubenswrapper[4838]: E1207 10:20:08.615076 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:20:19 crc kubenswrapper[4838]: I1207 10:20:19.615040 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:20:19 crc kubenswrapper[4838]: E1207 10:20:19.616919 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:20:33 crc kubenswrapper[4838]: I1207 10:20:33.620611 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:20:33 crc kubenswrapper[4838]: E1207 10:20:33.621509 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:20:39 crc kubenswrapper[4838]: I1207 10:20:39.791661 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59c4fb88bd-m7xbq_279b7609-3731-4b2d-a74f-64b591e53925/barbican-api/0.log" Dec 07 10:20:39 crc kubenswrapper[4838]: I1207 10:20:39.945217 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59c4fb88bd-m7xbq_279b7609-3731-4b2d-a74f-64b591e53925/barbican-api-log/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.039243 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d7849ff8d-sjw2m_2d766abd-30cf-45e4-9672-1f999276a187/barbican-keystone-listener/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.127979 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d7849ff8d-sjw2m_2d766abd-30cf-45e4-9672-1f999276a187/barbican-keystone-listener-log/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.283794 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-698f6d859c-2w95m_86c6d276-ba63-41ba-aa25-10cb0994f54a/barbican-worker/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.306075 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-698f6d859c-2w95m_86c6d276-ba63-41ba-aa25-10cb0994f54a/barbican-worker-log/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.430181 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd_1386eb32-0cef-4287-93d1-8fb5445081f3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.517850 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/ceilometer-central-agent/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.596945 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/ceilometer-notification-agent/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.642460 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/proxy-httpd/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.757771 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/sg-core/0.log" Dec 07 10:20:40 crc kubenswrapper[4838]: I1207 10:20:40.889394 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d_d681f081-977d-48c0-ac82-ff3c61647de8/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.014840 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56_9e6c35ea-c607-4c61-84d0-7bdf0d62add9/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.098395 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7e943cd3-c74f-4317-951f-5d1696ee7ffd/cinder-api/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.198280 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7e943cd3-c74f-4317-951f-5d1696ee7ffd/cinder-api-log/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.412331 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5d85f239-60d0-45b7-b884-a3b625160cc0/probe/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.480653 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5d85f239-60d0-45b7-b884-a3b625160cc0/cinder-backup/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.628783 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51d41156-ff39-4850-a562-5ba714cbd169/cinder-scheduler/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.720270 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51d41156-ff39-4850-a562-5ba714cbd169/probe/0.log" Dec 07 10:20:41 crc kubenswrapper[4838]: I1207 10:20:41.825740 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_874c3661-01ec-4c0c-8929-e05fc0fcf66e/cinder-volume/0.log" Dec 07 10:20:42 crc kubenswrapper[4838]: I1207 10:20:42.161896 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_874c3661-01ec-4c0c-8929-e05fc0fcf66e/probe/0.log" Dec 07 10:20:42 crc kubenswrapper[4838]: I1207 10:20:42.343166 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk_f5b4b264-d902-41a3-a100-6455a8a7b6ca/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:42 crc kubenswrapper[4838]: I1207 10:20:42.460527 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq_687f9d1b-115c-4b7a-bd3c-35666e665ec3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:42 crc kubenswrapper[4838]: I1207 10:20:42.617447 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7944d6f96c-r7hkg_2e347003-e564-4cf3-b006-9b4b3daf4044/init/0.log" Dec 07 10:20:42 crc kubenswrapper[4838]: I1207 10:20:42.893998 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7944d6f96c-r7hkg_2e347003-e564-4cf3-b006-9b4b3daf4044/dnsmasq-dns/0.log" Dec 07 10:20:42 crc kubenswrapper[4838]: I1207 10:20:42.902673 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f8f155e0-a710-40b0-af6b-157bb7e25854/glance-httpd/0.log" Dec 07 10:20:42 crc kubenswrapper[4838]: I1207 10:20:42.909666 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7944d6f96c-r7hkg_2e347003-e564-4cf3-b006-9b4b3daf4044/init/0.log" Dec 07 10:20:43 crc kubenswrapper[4838]: I1207 10:20:43.494840 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_57a396f6-7966-45b4-a4a5-4f91e688eaaa/glance-log/0.log" Dec 07 10:20:43 crc kubenswrapper[4838]: I1207 10:20:43.518054 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f8f155e0-a710-40b0-af6b-157bb7e25854/glance-log/0.log" Dec 07 10:20:43 crc kubenswrapper[4838]: I1207 10:20:43.564296 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_57a396f6-7966-45b4-a4a5-4f91e688eaaa/glance-httpd/0.log" Dec 07 10:20:43 crc kubenswrapper[4838]: I1207 10:20:43.821218 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-76f6479d76-n4xcx_3f6fe9f7-9f77-4da5-823b-b650957c5860/horizon/0.log" Dec 07 10:20:43 crc kubenswrapper[4838]: I1207 10:20:43.886287 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-76f6479d76-n4xcx_3f6fe9f7-9f77-4da5-823b-b650957c5860/horizon-log/0.log" Dec 07 10:20:44 crc kubenswrapper[4838]: I1207 10:20:44.311966 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s_84df8337-c690-4c33-9d82-da53b4de5be5/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:44 crc kubenswrapper[4838]: I1207 10:20:44.327441 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nlmxt_6b76e0a9-db0a-4913-90a7-1306215b82b7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:44 crc kubenswrapper[4838]: I1207 10:20:44.539945 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-654fd9b8c-czchm_dc215482-4e46-4b74-83da-82e55658e2d6/keystone-api/0.log" Dec 07 10:20:44 crc kubenswrapper[4838]: I1207 10:20:44.554579 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29418361-mnpmk_5de7aef2-84f7-4398-a557-4041f9031e31/keystone-cron/0.log" Dec 07 10:20:44 crc kubenswrapper[4838]: I1207 10:20:44.682860 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a10a8e53-204b-48d6-aaef-809d1847f251/kube-state-metrics/0.log" Dec 07 10:20:45 crc kubenswrapper[4838]: I1207 10:20:45.262006 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt_47331703-36e3-47e2-bbe0-e891130bdbe7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:45 crc kubenswrapper[4838]: I1207 10:20:45.262865 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_79c0afb9-cd0a-4608-87d1-1163282d784a/manila-api-log/0.log" Dec 07 10:20:45 crc kubenswrapper[4838]: I1207 10:20:45.337802 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_79c0afb9-cd0a-4608-87d1-1163282d784a/manila-api/0.log" Dec 07 10:20:45 crc kubenswrapper[4838]: I1207 10:20:45.476590 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_f404f222-0d07-479e-a0d9-0cb1ce143f3e/probe/0.log" Dec 07 10:20:45 crc kubenswrapper[4838]: I1207 10:20:45.535910 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_f404f222-0d07-479e-a0d9-0cb1ce143f3e/manila-scheduler/0.log" Dec 07 10:20:45 crc kubenswrapper[4838]: I1207 10:20:45.645493 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b5e3f331-938e-4ca9-842e-a580af7035d2/manila-share/0.log" Dec 07 10:20:45 crc kubenswrapper[4838]: I1207 10:20:45.702339 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b5e3f331-938e-4ca9-842e-a580af7035d2/probe/0.log" Dec 07 10:20:46 crc kubenswrapper[4838]: I1207 10:20:46.028035 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7998cf6c8f-xbbtt_562c9fcd-8827-42a8-8753-d570d77656fc/neutron-httpd/0.log" Dec 07 10:20:46 crc kubenswrapper[4838]: I1207 10:20:46.079599 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7998cf6c8f-xbbtt_562c9fcd-8827-42a8-8753-d570d77656fc/neutron-api/0.log" Dec 07 10:20:46 crc kubenswrapper[4838]: I1207 10:20:46.211580 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng_63630234-9950-43db-a3da-a4ea9ba22be3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:46 crc kubenswrapper[4838]: I1207 10:20:46.723954 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_485d8af5-c6b8-4d04-be66-a3a421f86d49/nova-cell0-conductor-conductor/0.log" Dec 07 10:20:46 crc kubenswrapper[4838]: I1207 10:20:46.904496 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1d429890-5092-448b-b386-36493df22443/nova-api-log/0.log" Dec 07 10:20:47 crc kubenswrapper[4838]: I1207 10:20:47.093552 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_8cdfa44f-0d15-4338-8a82-82b624b9ba6d/nova-cell1-conductor-conductor/0.log" Dec 07 10:20:47 crc kubenswrapper[4838]: I1207 10:20:47.126143 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1d429890-5092-448b-b386-36493df22443/nova-api-api/0.log" Dec 07 10:20:47 crc kubenswrapper[4838]: I1207 10:20:47.225341 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_cb3ea478-6a63-4e03-b9c1-890fca36b984/nova-cell1-novncproxy-novncproxy/0.log" Dec 07 10:20:47 crc kubenswrapper[4838]: I1207 10:20:47.428527 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd_e47e7bbd-a623-4a13-ba01-2fb7ff984b40/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:47 crc kubenswrapper[4838]: I1207 10:20:47.654154 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_021e33aa-1082-46ef-afa0-1dd04aad6502/nova-metadata-log/0.log" Dec 07 10:20:47 crc kubenswrapper[4838]: I1207 10:20:47.888635 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_0f3b4db3-4d04-4198-972d-ab3722b30abe/nova-scheduler-scheduler/0.log" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.041659 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c/mysql-bootstrap/0.log" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.147226 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c/mysql-bootstrap/0.log" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.366660 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c/galera/0.log" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.508649 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d111910-1e1a-4b37-87c2-5c8f55110a73/mysql-bootstrap/0.log" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.616996 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:20:48 crc kubenswrapper[4838]: E1207 10:20:48.617183 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.695463 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d111910-1e1a-4b37-87c2-5c8f55110a73/mysql-bootstrap/0.log" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.711381 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d111910-1e1a-4b37-87c2-5c8f55110a73/galera/0.log" Dec 07 10:20:48 crc kubenswrapper[4838]: I1207 10:20:48.932554 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_096a9643-43b6-4f09-8be1-cd739058c12e/openstackclient/0.log" Dec 07 10:20:49 crc kubenswrapper[4838]: I1207 10:20:49.099589 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-j7t4b_e7b95347-0585-4016-a3fa-a17c6d9042a3/openstack-network-exporter/0.log" Dec 07 10:20:49 crc kubenswrapper[4838]: I1207 10:20:49.194297 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_021e33aa-1082-46ef-afa0-1dd04aad6502/nova-metadata-metadata/0.log" Dec 07 10:20:49 crc kubenswrapper[4838]: I1207 10:20:49.331305 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovsdb-server-init/0.log" Dec 07 10:20:49 crc kubenswrapper[4838]: I1207 10:20:49.683620 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovs-vswitchd/0.log" Dec 07 10:20:49 crc kubenswrapper[4838]: I1207 10:20:49.705988 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovsdb-server/0.log" Dec 07 10:20:49 crc kubenswrapper[4838]: I1207 10:20:49.709995 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovsdb-server-init/0.log" Dec 07 10:20:49 crc kubenswrapper[4838]: I1207 10:20:49.853435 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-sbl7t_a3ae58f2-6cd2-4782-96b1-a88597b25b37/ovn-controller/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.074221 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-84r6c_6ae7a9e2-889e-499b-8f64-d38976e8d429/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.113029 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_065bed0f-e9fc-4d9a-a455-46d9d77e7bb2/openstack-network-exporter/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.258102 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_065bed0f-e9fc-4d9a-a455-46d9d77e7bb2/ovn-northd/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.390270 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cec74e7b-a2aa-46a0-b318-61decd213728/openstack-network-exporter/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.432929 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cec74e7b-a2aa-46a0-b318-61decd213728/ovsdbserver-nb/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.613470 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e06fb79a-09b7-42af-bb04-1ef9ba3f7560/openstack-network-exporter/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.698060 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e06fb79a-09b7-42af-bb04-1ef9ba3f7560/ovsdbserver-sb/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.849692 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9b574bc66-nsc2v_87509a0e-d4f9-4d24-b490-9d5e2675aa21/placement-api/0.log" Dec 07 10:20:50 crc kubenswrapper[4838]: I1207 10:20:50.982306 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a017a788-cb2e-45ed-bceb-26ea6b69a02d/setup-container/0.log" Dec 07 10:20:51 crc kubenswrapper[4838]: I1207 10:20:51.034186 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9b574bc66-nsc2v_87509a0e-d4f9-4d24-b490-9d5e2675aa21/placement-log/0.log" Dec 07 10:20:51 crc kubenswrapper[4838]: I1207 10:20:51.180274 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a017a788-cb2e-45ed-bceb-26ea6b69a02d/setup-container/0.log" Dec 07 10:20:51 crc kubenswrapper[4838]: I1207 10:20:51.241349 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a017a788-cb2e-45ed-bceb-26ea6b69a02d/rabbitmq/0.log" Dec 07 10:20:51 crc kubenswrapper[4838]: I1207 10:20:51.372697 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_688c3086-fd09-4b9d-921e-e41c7f8a01ac/setup-container/0.log" Dec 07 10:20:51 crc kubenswrapper[4838]: I1207 10:20:51.516843 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_688c3086-fd09-4b9d-921e-e41c7f8a01ac/setup-container/0.log" Dec 07 10:20:51 crc kubenswrapper[4838]: I1207 10:20:51.645808 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_688c3086-fd09-4b9d-921e-e41c7f8a01ac/rabbitmq/0.log" Dec 07 10:20:51 crc kubenswrapper[4838]: I1207 10:20:51.654779 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g_df0eb7df-51e9-4da6-9e69-0b11a79ad3fa/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:52 crc kubenswrapper[4838]: I1207 10:20:52.175829 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb_8587abfa-6182-4613-b458-69c4f1bf2eca/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:52 crc kubenswrapper[4838]: I1207 10:20:52.270010 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-d67k5_e471497e-008f-40a2-929e-33e08ea45f63/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:20:52 crc kubenswrapper[4838]: I1207 10:20:52.430331 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8grpd_131b49fd-19d7-4188-a131-2d684a0eee60/ssh-known-hosts-edpm-deployment/0.log" Dec 07 10:20:52 crc kubenswrapper[4838]: I1207 10:20:52.578027 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a430a791-ad62-4246-8422-6bae0b6ce772/tempest-tests-tempest-tests-runner/0.log" Dec 07 10:20:52 crc kubenswrapper[4838]: I1207 10:20:52.643533 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_17d4b7b1-e69f-444b-bd80-753b07d66af7/test-operator-logs-container/0.log" Dec 07 10:20:52 crc kubenswrapper[4838]: I1207 10:20:52.815510 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-99vbk_f9c17708-e8da-45b1-b919-354138adf40b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:21:00 crc kubenswrapper[4838]: I1207 10:21:00.614262 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:21:00 crc kubenswrapper[4838]: E1207 10:21:00.615020 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:21:05 crc kubenswrapper[4838]: I1207 10:21:05.878197 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_76cc3a82-a620-435a-99f0-9a91297d2ee2/memcached/0.log" Dec 07 10:21:14 crc kubenswrapper[4838]: I1207 10:21:14.614713 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:21:14 crc kubenswrapper[4838]: E1207 10:21:14.615548 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.076627 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/util/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.258059 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/util/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.258733 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/pull/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.320569 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/pull/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.531998 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/extract/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.541994 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/util/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.569230 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/pull/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.710866 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5cgxx_2e6976c0-74c7-4f62-88e0-6f6b635f829b/kube-rbac-proxy/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.806845 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5cgxx_2e6976c0-74c7-4f62-88e0-6f6b635f829b/manager/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.833572 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-p5xdf_e9af8ce9-a7b0-4bca-b834-2f2f494eb434/kube-rbac-proxy/0.log" Dec 07 10:21:25 crc kubenswrapper[4838]: I1207 10:21:25.990246 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-p5xdf_e9af8ce9-a7b0-4bca-b834-2f2f494eb434/manager/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.048386 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vb8wb_c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd/kube-rbac-proxy/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.084724 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vb8wb_c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd/manager/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.259065 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-9hg7c_f91ec7b3-50b5-4f9f-95a0-c65abeeb879f/kube-rbac-proxy/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.296071 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-9hg7c_f91ec7b3-50b5-4f9f-95a0-c65abeeb879f/manager/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.445796 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4hn7r_5d086115-289a-4478-a657-7ea3927165cd/kube-rbac-proxy/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.469553 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4hn7r_5d086115-289a-4478-a657-7ea3927165cd/manager/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.567366 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2ftqb_49a96bc4-d983-4088-973e-ebb95fdcc725/kube-rbac-proxy/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.614455 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:21:26 crc kubenswrapper[4838]: E1207 10:21:26.614790 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.657265 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2ftqb_49a96bc4-d983-4088-973e-ebb95fdcc725/manager/0.log" Dec 07 10:21:26 crc kubenswrapper[4838]: I1207 10:21:26.800644 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4q7pd_fac41ae4-04ea-4e07-b672-13fdf35a6530/kube-rbac-proxy/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.072124 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-pmdl9_ba02a228-91ff-41cf-88ac-b027e697dff9/manager/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.084842 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4q7pd_fac41ae4-04ea-4e07-b672-13fdf35a6530/manager/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.099778 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-pmdl9_ba02a228-91ff-41cf-88ac-b027e697dff9/kube-rbac-proxy/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.262539 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-gqrds_23f05536-d749-43c5-a7cc-7dc63f46e288/kube-rbac-proxy/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.362076 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-gqrds_23f05536-d749-43c5-a7cc-7dc63f46e288/manager/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.430958 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-hbffh_7feb13b5-ed7a-47bf-82f1-b749d5b9574e/kube-rbac-proxy/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.561687 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-hbffh_7feb13b5-ed7a-47bf-82f1-b749d5b9574e/manager/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.878067 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6cwcf_3bd0361d-590b-4a06-b32c-6f7ceb427fbd/kube-rbac-proxy/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.938962 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6cwcf_3bd0361d-590b-4a06-b32c-6f7ceb427fbd/manager/0.log" Dec 07 10:21:27 crc kubenswrapper[4838]: I1207 10:21:27.978931 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-9cwrh_3af86462-0c95-4f05-9e30-c70da589a944/kube-rbac-proxy/0.log" Dec 07 10:21:28 crc kubenswrapper[4838]: I1207 10:21:28.191438 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-t6tz7_86196566-5b23-4cf1-924e-8a802086e443/kube-rbac-proxy/0.log" Dec 07 10:21:28 crc kubenswrapper[4838]: I1207 10:21:28.218706 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-9cwrh_3af86462-0c95-4f05-9e30-c70da589a944/manager/0.log" Dec 07 10:21:28 crc kubenswrapper[4838]: I1207 10:21:28.354505 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-t6tz7_86196566-5b23-4cf1-924e-8a802086e443/manager/0.log" Dec 07 10:21:28 crc kubenswrapper[4838]: I1207 10:21:28.419709 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-54jxx_30216693-f616-448b-b85d-1c6482317ec5/kube-rbac-proxy/0.log" Dec 07 10:21:28 crc kubenswrapper[4838]: I1207 10:21:28.494309 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-54jxx_30216693-f616-448b-b85d-1c6482317ec5/manager/0.log" Dec 07 10:21:28 crc kubenswrapper[4838]: I1207 10:21:28.660319 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-744f8cb766qxndm_07975bb2-d979-4265-bba2-1254d58b7267/kube-rbac-proxy/0.log" Dec 07 10:21:28 crc kubenswrapper[4838]: I1207 10:21:28.823478 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-744f8cb766qxndm_07975bb2-d979-4265-bba2-1254d58b7267/manager/0.log" Dec 07 10:21:29 crc kubenswrapper[4838]: I1207 10:21:29.213064 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5f48db4cb9-cs44h_e28de98c-d75e-4673-8299-fd5abba22b44/operator/0.log" Dec 07 10:21:29 crc kubenswrapper[4838]: I1207 10:21:29.755740 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-b29lr_fdc08e45-9f41-4a0c-8478-10ab2b22ad4e/kube-rbac-proxy/0.log" Dec 07 10:21:29 crc kubenswrapper[4838]: I1207 10:21:29.760141 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-b29lr_fdc08e45-9f41-4a0c-8478-10ab2b22ad4e/manager/0.log" Dec 07 10:21:29 crc kubenswrapper[4838]: I1207 10:21:29.828543 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2sblj_0851e8d2-5d7c-4e07-aed5-d743468017ec/registry-server/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.058117 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9jtsq_5fc12298-865b-411d-ab04-6cc9d99d8030/kube-rbac-proxy/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.119026 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9jtsq_5fc12298-865b-411d-ab04-6cc9d99d8030/manager/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.140370 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5b47684954-7mp9q_1f077e27-60ec-4c93-aada-4a002c07a70c/manager/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.203448 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-tl2fn_700b9dd5-3c10-4fa7-bde4-0c9a4f778608/operator/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.298892 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-6vrx6_5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f/kube-rbac-proxy/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.344120 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-6vrx6_5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f/manager/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.397446 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-gbbg9_71940e0c-91e5-4634-92d3-d6061fbfc768/kube-rbac-proxy/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.537503 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-gbbg9_71940e0c-91e5-4634-92d3-d6061fbfc768/manager/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.587695 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wz9hz_44803609-79be-4c16-bd10-3d6820d8f24b/kube-rbac-proxy/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.654676 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wz9hz_44803609-79be-4c16-bd10-3d6820d8f24b/manager/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.718908 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-v5hzl_3c038260-fe42-4c67-8e8c-b994a83c6156/kube-rbac-proxy/0.log" Dec 07 10:21:30 crc kubenswrapper[4838]: I1207 10:21:30.767387 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-v5hzl_3c038260-fe42-4c67-8e8c-b994a83c6156/manager/0.log" Dec 07 10:21:40 crc kubenswrapper[4838]: I1207 10:21:40.614901 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:21:40 crc kubenswrapper[4838]: E1207 10:21:40.615633 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:21:51 crc kubenswrapper[4838]: I1207 10:21:51.614981 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:21:51 crc kubenswrapper[4838]: E1207 10:21:51.615639 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:21:52 crc kubenswrapper[4838]: I1207 10:21:52.838948 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pcrsf_22165712-08e9-47cc-aaab-e5da9f92a0fb/control-plane-machine-set-operator/0.log" Dec 07 10:21:53 crc kubenswrapper[4838]: I1207 10:21:53.016206 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qll5w_e432ea72-cff7-4826-8001-125443ca158a/kube-rbac-proxy/0.log" Dec 07 10:21:53 crc kubenswrapper[4838]: I1207 10:21:53.082388 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qll5w_e432ea72-cff7-4826-8001-125443ca158a/machine-api-operator/0.log" Dec 07 10:22:02 crc kubenswrapper[4838]: I1207 10:22:02.614463 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:22:02 crc kubenswrapper[4838]: E1207 10:22:02.615316 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:22:07 crc kubenswrapper[4838]: I1207 10:22:07.968154 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2h6mq_d7e66c54-9769-4287-8b1c-f1d1a56b08a1/cert-manager-controller/0.log" Dec 07 10:22:08 crc kubenswrapper[4838]: I1207 10:22:08.023186 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-6knh9_c0fb9b54-1b0a-4cab-8bad-669c11ea8a73/cert-manager-cainjector/0.log" Dec 07 10:22:08 crc kubenswrapper[4838]: I1207 10:22:08.187004 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-87mkv_72c34e1b-8df3-4622-9cdb-b44adf0e45bf/cert-manager-webhook/0.log" Dec 07 10:22:13 crc kubenswrapper[4838]: I1207 10:22:13.639960 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:22:13 crc kubenswrapper[4838]: E1207 10:22:13.640764 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:22:21 crc kubenswrapper[4838]: I1207 10:22:21.331964 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-lzwd2_013a32de-8cd7-44da-b292-5948ab0bb6c7/nmstate-console-plugin/0.log" Dec 07 10:22:21 crc kubenswrapper[4838]: I1207 10:22:21.392952 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-twb76_a130a77c-66e9-48b2-a9c9-36c44b7b8eec/nmstate-handler/0.log" Dec 07 10:22:21 crc kubenswrapper[4838]: I1207 10:22:21.585590 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p58hh_11834459-a487-4c2f-9521-add0e1726e9d/kube-rbac-proxy/0.log" Dec 07 10:22:21 crc kubenswrapper[4838]: I1207 10:22:21.639236 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p58hh_11834459-a487-4c2f-9521-add0e1726e9d/nmstate-metrics/0.log" Dec 07 10:22:21 crc kubenswrapper[4838]: I1207 10:22:21.748352 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-4hcdm_e1df0bd8-ff98-4094-92b6-23f266883f8a/nmstate-operator/0.log" Dec 07 10:22:21 crc kubenswrapper[4838]: I1207 10:22:21.822136 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-jcgsg_9d7f7de9-0d6e-4bf5-83c0-5f703cb433af/nmstate-webhook/0.log" Dec 07 10:22:24 crc kubenswrapper[4838]: I1207 10:22:24.615455 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:22:24 crc kubenswrapper[4838]: E1207 10:22:24.616119 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:22:35 crc kubenswrapper[4838]: I1207 10:22:35.618011 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:22:35 crc kubenswrapper[4838]: E1207 10:22:35.619237 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.285055 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj79s_4fa3fe39-d052-4d6f-881c-526b036b4bbf/controller/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.289379 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj79s_4fa3fe39-d052-4d6f-881c-526b036b4bbf/kube-rbac-proxy/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.314251 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-6btm8_39164f82-bb66-4fe1-a4e4-0007d9697e40/frr-k8s-webhook-server/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.521735 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.774541 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.783672 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.806674 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.837363 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:22:38 crc kubenswrapper[4838]: I1207 10:22:38.967646 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.030292 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.078279 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.126166 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.456397 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.466409 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.527401 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.592482 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/controller/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.674307 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/frr-metrics/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.735282 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/kube-rbac-proxy/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.858830 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/reloader/0.log" Dec 07 10:22:39 crc kubenswrapper[4838]: I1207 10:22:39.912494 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/kube-rbac-proxy-frr/0.log" Dec 07 10:22:40 crc kubenswrapper[4838]: I1207 10:22:40.088806 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d7fb758f7-hm2jk_53c25e7b-09e8-435d-933b-9cbbe46f5ac0/manager/0.log" Dec 07 10:22:40 crc kubenswrapper[4838]: I1207 10:22:40.303718 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-94cfbc768-26mzl_cda6c003-8725-4f92-a8a8-6dad25837d56/webhook-server/0.log" Dec 07 10:22:40 crc kubenswrapper[4838]: I1207 10:22:40.461929 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vdgqx_a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86/kube-rbac-proxy/0.log" Dec 07 10:22:41 crc kubenswrapper[4838]: I1207 10:22:41.086842 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vdgqx_a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86/speaker/0.log" Dec 07 10:22:41 crc kubenswrapper[4838]: I1207 10:22:41.115704 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/frr/0.log" Dec 07 10:22:49 crc kubenswrapper[4838]: I1207 10:22:49.615285 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:22:49 crc kubenswrapper[4838]: E1207 10:22:49.616254 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:22:55 crc kubenswrapper[4838]: I1207 10:22:55.559985 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/util/0.log" Dec 07 10:22:55 crc kubenswrapper[4838]: I1207 10:22:55.748321 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/pull/0.log" Dec 07 10:22:55 crc kubenswrapper[4838]: I1207 10:22:55.813101 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/util/0.log" Dec 07 10:22:55 crc kubenswrapper[4838]: I1207 10:22:55.816180 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/pull/0.log" Dec 07 10:22:55 crc kubenswrapper[4838]: I1207 10:22:55.953727 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/util/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.011544 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/pull/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.026778 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/extract/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.161097 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/util/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.320541 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/pull/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.361932 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/util/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.385494 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/pull/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.563664 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/pull/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.568219 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/util/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.586209 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/extract/0.log" Dec 07 10:22:56 crc kubenswrapper[4838]: I1207 10:22:56.760099 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-utilities/0.log" Dec 07 10:22:57 crc kubenswrapper[4838]: I1207 10:22:57.285207 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-content/0.log" Dec 07 10:22:57 crc kubenswrapper[4838]: I1207 10:22:57.352480 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-content/0.log" Dec 07 10:22:57 crc kubenswrapper[4838]: I1207 10:22:57.374235 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-utilities/0.log" Dec 07 10:22:57 crc kubenswrapper[4838]: I1207 10:22:57.542865 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-utilities/0.log" Dec 07 10:22:57 crc kubenswrapper[4838]: I1207 10:22:57.577076 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-content/0.log" Dec 07 10:22:57 crc kubenswrapper[4838]: I1207 10:22:57.934300 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-utilities/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.057157 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/registry-server/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.181131 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-utilities/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.204371 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-content/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.204372 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-content/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.396679 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-utilities/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.412898 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-content/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.791252 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-69xn8_c21136ea-de51-48f3-b79f-493f2d88ece5/marketplace-operator/0.log" Dec 07 10:22:58 crc kubenswrapper[4838]: I1207 10:22:58.824539 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-utilities/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.091552 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/registry-server/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.208875 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-content/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.220523 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-content/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.276654 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-utilities/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.435191 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-utilities/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.463465 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-content/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.572519 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-utilities/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.591506 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/registry-server/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.746548 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-utilities/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.767331 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-content/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.786747 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-content/0.log" Dec 07 10:22:59 crc kubenswrapper[4838]: I1207 10:22:59.998538 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-content/0.log" Dec 07 10:23:00 crc kubenswrapper[4838]: I1207 10:23:00.019178 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-utilities/0.log" Dec 07 10:23:00 crc kubenswrapper[4838]: I1207 10:23:00.475675 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/registry-server/0.log" Dec 07 10:23:02 crc kubenswrapper[4838]: I1207 10:23:02.614126 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.062646 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"11045922bb917c10a0c80177be4aee6408cbd108c65ad5b2f6027c8530522657"} Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.345706 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hqnht"] Dec 07 10:23:03 crc kubenswrapper[4838]: E1207 10:23:03.346664 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6263c5b5-1a1e-4840-b3a1-2241162c05d9" containerName="container-00" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.346690 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="6263c5b5-1a1e-4840-b3a1-2241162c05d9" containerName="container-00" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.346947 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="6263c5b5-1a1e-4840-b3a1-2241162c05d9" containerName="container-00" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.348539 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.365097 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hqnht"] Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.399981 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-utilities\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.400083 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6qvw\" (UniqueName: \"kubernetes.io/projected/c867de93-9253-4f6e-8b1b-2bdb63981c53-kube-api-access-w6qvw\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.400160 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-catalog-content\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.510444 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6qvw\" (UniqueName: \"kubernetes.io/projected/c867de93-9253-4f6e-8b1b-2bdb63981c53-kube-api-access-w6qvw\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.510509 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-catalog-content\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.510848 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-utilities\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.511287 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-utilities\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.511776 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-catalog-content\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.562677 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6qvw\" (UniqueName: \"kubernetes.io/projected/c867de93-9253-4f6e-8b1b-2bdb63981c53-kube-api-access-w6qvw\") pod \"redhat-marketplace-hqnht\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:03 crc kubenswrapper[4838]: I1207 10:23:03.678076 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:04 crc kubenswrapper[4838]: I1207 10:23:04.269641 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hqnht"] Dec 07 10:23:05 crc kubenswrapper[4838]: I1207 10:23:05.092692 4838 generic.go:334] "Generic (PLEG): container finished" podID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerID="7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135" exitCode=0 Dec 07 10:23:05 crc kubenswrapper[4838]: I1207 10:23:05.093187 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hqnht" event={"ID":"c867de93-9253-4f6e-8b1b-2bdb63981c53","Type":"ContainerDied","Data":"7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135"} Dec 07 10:23:05 crc kubenswrapper[4838]: I1207 10:23:05.093218 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hqnht" event={"ID":"c867de93-9253-4f6e-8b1b-2bdb63981c53","Type":"ContainerStarted","Data":"5660f34243ec80b8392cf95216fdb171fc9bda59d81bf030ce10c538878c871b"} Dec 07 10:23:05 crc kubenswrapper[4838]: I1207 10:23:05.096488 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.115006 4838 generic.go:334] "Generic (PLEG): container finished" podID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerID="9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8" exitCode=0 Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.115099 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hqnht" event={"ID":"c867de93-9253-4f6e-8b1b-2bdb63981c53","Type":"ContainerDied","Data":"9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8"} Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.170930 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rdx8x"] Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.175580 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.187435 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdx8x"] Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.285192 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-catalog-content\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.285349 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqf8v\" (UniqueName: \"kubernetes.io/projected/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-kube-api-access-mqf8v\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.285707 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-utilities\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.387166 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-catalog-content\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.387245 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqf8v\" (UniqueName: \"kubernetes.io/projected/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-kube-api-access-mqf8v\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.387762 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-catalog-content\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.387951 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-utilities\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.388300 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-utilities\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.409892 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqf8v\" (UniqueName: \"kubernetes.io/projected/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-kube-api-access-mqf8v\") pod \"redhat-operators-rdx8x\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:07 crc kubenswrapper[4838]: I1207 10:23:07.504097 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:08 crc kubenswrapper[4838]: I1207 10:23:08.051462 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdx8x"] Dec 07 10:23:08 crc kubenswrapper[4838]: W1207 10:23:08.052419 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podececbb1b_8f4a_4563_8c48_4b3cf57a6d62.slice/crio-870325ab6eeb51c498f0b3cca2d83a6ffc57bf3376a398bfc6fceacf343d15ad WatchSource:0}: Error finding container 870325ab6eeb51c498f0b3cca2d83a6ffc57bf3376a398bfc6fceacf343d15ad: Status 404 returned error can't find the container with id 870325ab6eeb51c498f0b3cca2d83a6ffc57bf3376a398bfc6fceacf343d15ad Dec 07 10:23:08 crc kubenswrapper[4838]: I1207 10:23:08.123836 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdx8x" event={"ID":"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62","Type":"ContainerStarted","Data":"870325ab6eeb51c498f0b3cca2d83a6ffc57bf3376a398bfc6fceacf343d15ad"} Dec 07 10:23:08 crc kubenswrapper[4838]: I1207 10:23:08.125670 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hqnht" event={"ID":"c867de93-9253-4f6e-8b1b-2bdb63981c53","Type":"ContainerStarted","Data":"0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd"} Dec 07 10:23:08 crc kubenswrapper[4838]: I1207 10:23:08.143905 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hqnht" podStartSLOduration=2.70126955 podStartE2EDuration="5.143886982s" podCreationTimestamp="2025-12-07 10:23:03 +0000 UTC" firstStartedPulling="2025-12-07 10:23:05.096240629 +0000 UTC m=+4601.803559656" lastFinishedPulling="2025-12-07 10:23:07.538858071 +0000 UTC m=+4604.246177088" observedRunningTime="2025-12-07 10:23:08.142265137 +0000 UTC m=+4604.849584154" watchObservedRunningTime="2025-12-07 10:23:08.143886982 +0000 UTC m=+4604.851205999" Dec 07 10:23:09 crc kubenswrapper[4838]: I1207 10:23:09.135935 4838 generic.go:334] "Generic (PLEG): container finished" podID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerID="d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983" exitCode=0 Dec 07 10:23:09 crc kubenswrapper[4838]: I1207 10:23:09.136036 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdx8x" event={"ID":"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62","Type":"ContainerDied","Data":"d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983"} Dec 07 10:23:10 crc kubenswrapper[4838]: I1207 10:23:10.146566 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdx8x" event={"ID":"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62","Type":"ContainerStarted","Data":"646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef"} Dec 07 10:23:13 crc kubenswrapper[4838]: I1207 10:23:13.679086 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:13 crc kubenswrapper[4838]: I1207 10:23:13.679601 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:13 crc kubenswrapper[4838]: I1207 10:23:13.981140 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:14 crc kubenswrapper[4838]: I1207 10:23:14.188643 4838 generic.go:334] "Generic (PLEG): container finished" podID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerID="646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef" exitCode=0 Dec 07 10:23:14 crc kubenswrapper[4838]: I1207 10:23:14.188933 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdx8x" event={"ID":"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62","Type":"ContainerDied","Data":"646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef"} Dec 07 10:23:14 crc kubenswrapper[4838]: I1207 10:23:14.264091 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:15 crc kubenswrapper[4838]: I1207 10:23:15.199477 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdx8x" event={"ID":"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62","Type":"ContainerStarted","Data":"191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8"} Dec 07 10:23:15 crc kubenswrapper[4838]: I1207 10:23:15.230041 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rdx8x" podStartSLOduration=2.721762051 podStartE2EDuration="8.230024941s" podCreationTimestamp="2025-12-07 10:23:07 +0000 UTC" firstStartedPulling="2025-12-07 10:23:09.13778806 +0000 UTC m=+4605.845107077" lastFinishedPulling="2025-12-07 10:23:14.64605095 +0000 UTC m=+4611.353369967" observedRunningTime="2025-12-07 10:23:15.221704987 +0000 UTC m=+4611.929024004" watchObservedRunningTime="2025-12-07 10:23:15.230024941 +0000 UTC m=+4611.937343958" Dec 07 10:23:16 crc kubenswrapper[4838]: I1207 10:23:16.504608 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hqnht"] Dec 07 10:23:16 crc kubenswrapper[4838]: I1207 10:23:16.505056 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hqnht" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="registry-server" containerID="cri-o://0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd" gracePeriod=2 Dec 07 10:23:16 crc kubenswrapper[4838]: I1207 10:23:16.975753 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.097446 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6qvw\" (UniqueName: \"kubernetes.io/projected/c867de93-9253-4f6e-8b1b-2bdb63981c53-kube-api-access-w6qvw\") pod \"c867de93-9253-4f6e-8b1b-2bdb63981c53\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.097520 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-catalog-content\") pod \"c867de93-9253-4f6e-8b1b-2bdb63981c53\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.097663 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-utilities\") pod \"c867de93-9253-4f6e-8b1b-2bdb63981c53\" (UID: \"c867de93-9253-4f6e-8b1b-2bdb63981c53\") " Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.098518 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-utilities" (OuterVolumeSpecName: "utilities") pod "c867de93-9253-4f6e-8b1b-2bdb63981c53" (UID: "c867de93-9253-4f6e-8b1b-2bdb63981c53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.116678 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c867de93-9253-4f6e-8b1b-2bdb63981c53" (UID: "c867de93-9253-4f6e-8b1b-2bdb63981c53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.128096 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c867de93-9253-4f6e-8b1b-2bdb63981c53-kube-api-access-w6qvw" (OuterVolumeSpecName: "kube-api-access-w6qvw") pod "c867de93-9253-4f6e-8b1b-2bdb63981c53" (UID: "c867de93-9253-4f6e-8b1b-2bdb63981c53"). InnerVolumeSpecName "kube-api-access-w6qvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.200452 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.200477 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6qvw\" (UniqueName: \"kubernetes.io/projected/c867de93-9253-4f6e-8b1b-2bdb63981c53-kube-api-access-w6qvw\") on node \"crc\" DevicePath \"\"" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.200487 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c867de93-9253-4f6e-8b1b-2bdb63981c53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.237394 4838 generic.go:334] "Generic (PLEG): container finished" podID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerID="0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd" exitCode=0 Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.237621 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hqnht" event={"ID":"c867de93-9253-4f6e-8b1b-2bdb63981c53","Type":"ContainerDied","Data":"0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd"} Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.237650 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hqnht" event={"ID":"c867de93-9253-4f6e-8b1b-2bdb63981c53","Type":"ContainerDied","Data":"5660f34243ec80b8392cf95216fdb171fc9bda59d81bf030ce10c538878c871b"} Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.237666 4838 scope.go:117] "RemoveContainer" containerID="0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.237791 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hqnht" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.256891 4838 scope.go:117] "RemoveContainer" containerID="9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.274915 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hqnht"] Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.285890 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hqnht"] Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.298362 4838 scope.go:117] "RemoveContainer" containerID="7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.339799 4838 scope.go:117] "RemoveContainer" containerID="0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd" Dec 07 10:23:17 crc kubenswrapper[4838]: E1207 10:23:17.340254 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd\": container with ID starting with 0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd not found: ID does not exist" containerID="0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.340281 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd"} err="failed to get container status \"0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd\": rpc error: code = NotFound desc = could not find container \"0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd\": container with ID starting with 0eae29ab33ab53ef724186c34c7f2c864cfe778415d3d96a544c7a586d26e3dd not found: ID does not exist" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.340301 4838 scope.go:117] "RemoveContainer" containerID="9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8" Dec 07 10:23:17 crc kubenswrapper[4838]: E1207 10:23:17.340516 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8\": container with ID starting with 9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8 not found: ID does not exist" containerID="9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.340538 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8"} err="failed to get container status \"9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8\": rpc error: code = NotFound desc = could not find container \"9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8\": container with ID starting with 9fcbcb3491ed9b3dd15401675699333c4ffa09d11e2cac731883608175ca55b8 not found: ID does not exist" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.340551 4838 scope.go:117] "RemoveContainer" containerID="7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135" Dec 07 10:23:17 crc kubenswrapper[4838]: E1207 10:23:17.341085 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135\": container with ID starting with 7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135 not found: ID does not exist" containerID="7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.341136 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135"} err="failed to get container status \"7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135\": rpc error: code = NotFound desc = could not find container \"7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135\": container with ID starting with 7cc54a0ee9f18f2dc47cc2c68450825061af0c09af8ed36e9b373ae391689135 not found: ID does not exist" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.504931 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.504983 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:17 crc kubenswrapper[4838]: I1207 10:23:17.624526 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" path="/var/lib/kubelet/pods/c867de93-9253-4f6e-8b1b-2bdb63981c53/volumes" Dec 07 10:23:18 crc kubenswrapper[4838]: I1207 10:23:18.581586 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rdx8x" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="registry-server" probeResult="failure" output=< Dec 07 10:23:18 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 10:23:18 crc kubenswrapper[4838]: > Dec 07 10:23:29 crc kubenswrapper[4838]: I1207 10:23:29.057990 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rdx8x" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="registry-server" probeResult="failure" output=< Dec 07 10:23:29 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 10:23:29 crc kubenswrapper[4838]: > Dec 07 10:23:37 crc kubenswrapper[4838]: I1207 10:23:37.548707 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:37 crc kubenswrapper[4838]: I1207 10:23:37.593899 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:38 crc kubenswrapper[4838]: I1207 10:23:38.331910 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdx8x"] Dec 07 10:23:39 crc kubenswrapper[4838]: I1207 10:23:39.419404 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rdx8x" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="registry-server" containerID="cri-o://191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8" gracePeriod=2 Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.154332 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.268270 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-utilities\") pod \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.268642 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-catalog-content\") pod \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.268666 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqf8v\" (UniqueName: \"kubernetes.io/projected/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-kube-api-access-mqf8v\") pod \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\" (UID: \"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62\") " Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.270216 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-utilities" (OuterVolumeSpecName: "utilities") pod "ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" (UID: "ececbb1b-8f4a-4563-8c48-4b3cf57a6d62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.277084 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-kube-api-access-mqf8v" (OuterVolumeSpecName: "kube-api-access-mqf8v") pod "ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" (UID: "ececbb1b-8f4a-4563-8c48-4b3cf57a6d62"). InnerVolumeSpecName "kube-api-access-mqf8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.371912 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqf8v\" (UniqueName: \"kubernetes.io/projected/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-kube-api-access-mqf8v\") on node \"crc\" DevicePath \"\"" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.371941 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.425619 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" (UID: "ececbb1b-8f4a-4563-8c48-4b3cf57a6d62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.432769 4838 generic.go:334] "Generic (PLEG): container finished" podID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerID="191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8" exitCode=0 Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.432804 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdx8x" event={"ID":"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62","Type":"ContainerDied","Data":"191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8"} Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.432902 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdx8x" event={"ID":"ececbb1b-8f4a-4563-8c48-4b3cf57a6d62","Type":"ContainerDied","Data":"870325ab6eeb51c498f0b3cca2d83a6ffc57bf3376a398bfc6fceacf343d15ad"} Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.432918 4838 scope.go:117] "RemoveContainer" containerID="191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.433022 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdx8x" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.465246 4838 scope.go:117] "RemoveContainer" containerID="646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.473966 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdx8x"] Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.475118 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.489640 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rdx8x"] Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.526435 4838 scope.go:117] "RemoveContainer" containerID="d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.543217 4838 scope.go:117] "RemoveContainer" containerID="191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8" Dec 07 10:23:40 crc kubenswrapper[4838]: E1207 10:23:40.543546 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8\": container with ID starting with 191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8 not found: ID does not exist" containerID="191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.543617 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8"} err="failed to get container status \"191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8\": rpc error: code = NotFound desc = could not find container \"191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8\": container with ID starting with 191d75b568866049103ef8f8a43a1cc4945a6e0ecbfb28a09d1122a8972213f8 not found: ID does not exist" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.543638 4838 scope.go:117] "RemoveContainer" containerID="646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef" Dec 07 10:23:40 crc kubenswrapper[4838]: E1207 10:23:40.543849 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef\": container with ID starting with 646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef not found: ID does not exist" containerID="646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.543872 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef"} err="failed to get container status \"646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef\": rpc error: code = NotFound desc = could not find container \"646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef\": container with ID starting with 646715fe6f1ff59baf45f54ce61099428c26ff2d08c9e631143301d8fb3d82ef not found: ID does not exist" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.543888 4838 scope.go:117] "RemoveContainer" containerID="d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983" Dec 07 10:23:40 crc kubenswrapper[4838]: E1207 10:23:40.544121 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983\": container with ID starting with d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983 not found: ID does not exist" containerID="d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983" Dec 07 10:23:40 crc kubenswrapper[4838]: I1207 10:23:40.544141 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983"} err="failed to get container status \"d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983\": rpc error: code = NotFound desc = could not find container \"d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983\": container with ID starting with d5c2c8d8de7cf0c94b82fb2c0e6ec1585aeeb6be45de69d753316d91542df983 not found: ID does not exist" Dec 07 10:23:41 crc kubenswrapper[4838]: I1207 10:23:41.623969 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" path="/var/lib/kubelet/pods/ececbb1b-8f4a-4563-8c48-4b3cf57a6d62/volumes" Dec 07 10:25:20 crc kubenswrapper[4838]: I1207 10:25:20.511152 4838 generic.go:334] "Generic (PLEG): container finished" podID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerID="46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633" exitCode=0 Dec 07 10:25:20 crc kubenswrapper[4838]: I1207 10:25:20.511215 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tjr69/must-gather-sqwk6" event={"ID":"fa0a6d32-e204-43ce-b6a0-609b93bea035","Type":"ContainerDied","Data":"46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633"} Dec 07 10:25:20 crc kubenswrapper[4838]: I1207 10:25:20.512242 4838 scope.go:117] "RemoveContainer" containerID="46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633" Dec 07 10:25:21 crc kubenswrapper[4838]: I1207 10:25:21.144132 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tjr69_must-gather-sqwk6_fa0a6d32-e204-43ce-b6a0-609b93bea035/gather/0.log" Dec 07 10:25:24 crc kubenswrapper[4838]: I1207 10:25:24.493647 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:25:24 crc kubenswrapper[4838]: I1207 10:25:24.494373 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:25:29 crc kubenswrapper[4838]: I1207 10:25:29.683563 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tjr69/must-gather-sqwk6"] Dec 07 10:25:29 crc kubenswrapper[4838]: I1207 10:25:29.684658 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-tjr69/must-gather-sqwk6" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerName="copy" containerID="cri-o://7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c" gracePeriod=2 Dec 07 10:25:29 crc kubenswrapper[4838]: I1207 10:25:29.693899 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tjr69/must-gather-sqwk6"] Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.190619 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tjr69_must-gather-sqwk6_fa0a6d32-e204-43ce-b6a0-609b93bea035/copy/0.log" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.191227 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.283669 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa0a6d32-e204-43ce-b6a0-609b93bea035-must-gather-output\") pod \"fa0a6d32-e204-43ce-b6a0-609b93bea035\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.284091 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv447\" (UniqueName: \"kubernetes.io/projected/fa0a6d32-e204-43ce-b6a0-609b93bea035-kube-api-access-cv447\") pod \"fa0a6d32-e204-43ce-b6a0-609b93bea035\" (UID: \"fa0a6d32-e204-43ce-b6a0-609b93bea035\") " Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.303241 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0a6d32-e204-43ce-b6a0-609b93bea035-kube-api-access-cv447" (OuterVolumeSpecName: "kube-api-access-cv447") pod "fa0a6d32-e204-43ce-b6a0-609b93bea035" (UID: "fa0a6d32-e204-43ce-b6a0-609b93bea035"). InnerVolumeSpecName "kube-api-access-cv447". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.386967 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv447\" (UniqueName: \"kubernetes.io/projected/fa0a6d32-e204-43ce-b6a0-609b93bea035-kube-api-access-cv447\") on node \"crc\" DevicePath \"\"" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.477333 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0a6d32-e204-43ce-b6a0-609b93bea035-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fa0a6d32-e204-43ce-b6a0-609b93bea035" (UID: "fa0a6d32-e204-43ce-b6a0-609b93bea035"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.488497 4838 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa0a6d32-e204-43ce-b6a0-609b93bea035-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.606463 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tjr69_must-gather-sqwk6_fa0a6d32-e204-43ce-b6a0-609b93bea035/copy/0.log" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.607215 4838 generic.go:334] "Generic (PLEG): container finished" podID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerID="7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c" exitCode=143 Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.607274 4838 scope.go:117] "RemoveContainer" containerID="7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.607452 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tjr69/must-gather-sqwk6" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.655957 4838 scope.go:117] "RemoveContainer" containerID="46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.724341 4838 scope.go:117] "RemoveContainer" containerID="7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c" Dec 07 10:25:30 crc kubenswrapper[4838]: E1207 10:25:30.725905 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c\": container with ID starting with 7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c not found: ID does not exist" containerID="7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.725960 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c"} err="failed to get container status \"7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c\": rpc error: code = NotFound desc = could not find container \"7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c\": container with ID starting with 7013cd70a1c6682b16bde962db3563d16b19d9f8f49824a815349f62198cf37c not found: ID does not exist" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.725986 4838 scope.go:117] "RemoveContainer" containerID="46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633" Dec 07 10:25:30 crc kubenswrapper[4838]: E1207 10:25:30.726343 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633\": container with ID starting with 46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633 not found: ID does not exist" containerID="46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633" Dec 07 10:25:30 crc kubenswrapper[4838]: I1207 10:25:30.726385 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633"} err="failed to get container status \"46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633\": rpc error: code = NotFound desc = could not find container \"46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633\": container with ID starting with 46ec22ac8af49e8416a68cdbc882c8aa635d202c59445a271f8e79f3d6e05633 not found: ID does not exist" Dec 07 10:25:31 crc kubenswrapper[4838]: I1207 10:25:31.632746 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" path="/var/lib/kubelet/pods/fa0a6d32-e204-43ce-b6a0-609b93bea035/volumes" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.273198 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b45m6"] Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.275206 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerName="gather" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.275329 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerName="gather" Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.275410 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="extract-content" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.275466 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="extract-content" Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.275525 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="extract-content" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.275579 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="extract-content" Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.275638 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="registry-server" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.275689 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="registry-server" Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.275757 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="registry-server" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.275808 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="registry-server" Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.275886 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerName="copy" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.275949 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerName="copy" Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.276014 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="extract-utilities" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.276066 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="extract-utilities" Dec 07 10:25:33 crc kubenswrapper[4838]: E1207 10:25:33.276128 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="extract-utilities" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.276177 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="extract-utilities" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.276398 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerName="copy" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.276474 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0a6d32-e204-43ce-b6a0-609b93bea035" containerName="gather" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.276532 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="ececbb1b-8f4a-4563-8c48-4b3cf57a6d62" containerName="registry-server" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.276592 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="c867de93-9253-4f6e-8b1b-2bdb63981c53" containerName="registry-server" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.283309 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.302538 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b45m6"] Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.388090 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvv67\" (UniqueName: \"kubernetes.io/projected/4cdbb8bb-a898-4674-8371-b85ad1743a1d-kube-api-access-bvv67\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.388255 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-catalog-content\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.388291 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-utilities\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.489962 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvv67\" (UniqueName: \"kubernetes.io/projected/4cdbb8bb-a898-4674-8371-b85ad1743a1d-kube-api-access-bvv67\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.490173 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-catalog-content\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.490217 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-utilities\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.490956 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-utilities\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.491047 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-catalog-content\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.518394 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvv67\" (UniqueName: \"kubernetes.io/projected/4cdbb8bb-a898-4674-8371-b85ad1743a1d-kube-api-access-bvv67\") pod \"certified-operators-b45m6\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.605912 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:33 crc kubenswrapper[4838]: I1207 10:25:33.865839 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b45m6"] Dec 07 10:25:34 crc kubenswrapper[4838]: I1207 10:25:34.651968 4838 generic.go:334] "Generic (PLEG): container finished" podID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerID="2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793" exitCode=0 Dec 07 10:25:34 crc kubenswrapper[4838]: I1207 10:25:34.652122 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b45m6" event={"ID":"4cdbb8bb-a898-4674-8371-b85ad1743a1d","Type":"ContainerDied","Data":"2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793"} Dec 07 10:25:34 crc kubenswrapper[4838]: I1207 10:25:34.676227 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b45m6" event={"ID":"4cdbb8bb-a898-4674-8371-b85ad1743a1d","Type":"ContainerStarted","Data":"3d1d270ed6e58741e94640c8cc70779811a2aff0de916fbe5217f05021f3b273"} Dec 07 10:25:35 crc kubenswrapper[4838]: I1207 10:25:35.686557 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b45m6" event={"ID":"4cdbb8bb-a898-4674-8371-b85ad1743a1d","Type":"ContainerStarted","Data":"b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1"} Dec 07 10:25:36 crc kubenswrapper[4838]: I1207 10:25:36.710089 4838 generic.go:334] "Generic (PLEG): container finished" podID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerID="b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1" exitCode=0 Dec 07 10:25:36 crc kubenswrapper[4838]: I1207 10:25:36.710170 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b45m6" event={"ID":"4cdbb8bb-a898-4674-8371-b85ad1743a1d","Type":"ContainerDied","Data":"b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1"} Dec 07 10:25:37 crc kubenswrapper[4838]: I1207 10:25:37.718992 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b45m6" event={"ID":"4cdbb8bb-a898-4674-8371-b85ad1743a1d","Type":"ContainerStarted","Data":"bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a"} Dec 07 10:25:37 crc kubenswrapper[4838]: I1207 10:25:37.749855 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b45m6" podStartSLOduration=2.321352511 podStartE2EDuration="4.749836495s" podCreationTimestamp="2025-12-07 10:25:33 +0000 UTC" firstStartedPulling="2025-12-07 10:25:34.655663056 +0000 UTC m=+4751.362982073" lastFinishedPulling="2025-12-07 10:25:37.084147 +0000 UTC m=+4753.791466057" observedRunningTime="2025-12-07 10:25:37.740668708 +0000 UTC m=+4754.447987735" watchObservedRunningTime="2025-12-07 10:25:37.749836495 +0000 UTC m=+4754.457155502" Dec 07 10:25:43 crc kubenswrapper[4838]: I1207 10:25:43.606554 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:43 crc kubenswrapper[4838]: I1207 10:25:43.607104 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:43 crc kubenswrapper[4838]: I1207 10:25:43.675026 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:43 crc kubenswrapper[4838]: I1207 10:25:43.820301 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:43 crc kubenswrapper[4838]: I1207 10:25:43.910170 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b45m6"] Dec 07 10:25:45 crc kubenswrapper[4838]: I1207 10:25:45.791704 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b45m6" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="registry-server" containerID="cri-o://bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a" gracePeriod=2 Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.280321 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.352660 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvv67\" (UniqueName: \"kubernetes.io/projected/4cdbb8bb-a898-4674-8371-b85ad1743a1d-kube-api-access-bvv67\") pod \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.352769 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-catalog-content\") pod \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.352918 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-utilities\") pod \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\" (UID: \"4cdbb8bb-a898-4674-8371-b85ad1743a1d\") " Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.355084 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-utilities" (OuterVolumeSpecName: "utilities") pod "4cdbb8bb-a898-4674-8371-b85ad1743a1d" (UID: "4cdbb8bb-a898-4674-8371-b85ad1743a1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.366016 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cdbb8bb-a898-4674-8371-b85ad1743a1d-kube-api-access-bvv67" (OuterVolumeSpecName: "kube-api-access-bvv67") pod "4cdbb8bb-a898-4674-8371-b85ad1743a1d" (UID: "4cdbb8bb-a898-4674-8371-b85ad1743a1d"). InnerVolumeSpecName "kube-api-access-bvv67". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.403560 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cdbb8bb-a898-4674-8371-b85ad1743a1d" (UID: "4cdbb8bb-a898-4674-8371-b85ad1743a1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.456090 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.456123 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvv67\" (UniqueName: \"kubernetes.io/projected/4cdbb8bb-a898-4674-8371-b85ad1743a1d-kube-api-access-bvv67\") on node \"crc\" DevicePath \"\"" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.456133 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cdbb8bb-a898-4674-8371-b85ad1743a1d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.805985 4838 generic.go:334] "Generic (PLEG): container finished" podID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerID="bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a" exitCode=0 Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.806048 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b45m6" event={"ID":"4cdbb8bb-a898-4674-8371-b85ad1743a1d","Type":"ContainerDied","Data":"bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a"} Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.806086 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b45m6" event={"ID":"4cdbb8bb-a898-4674-8371-b85ad1743a1d","Type":"ContainerDied","Data":"3d1d270ed6e58741e94640c8cc70779811a2aff0de916fbe5217f05021f3b273"} Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.806112 4838 scope.go:117] "RemoveContainer" containerID="bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.806264 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b45m6" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.837159 4838 scope.go:117] "RemoveContainer" containerID="b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.845278 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b45m6"] Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.855770 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b45m6"] Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.868900 4838 scope.go:117] "RemoveContainer" containerID="2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.907771 4838 scope.go:117] "RemoveContainer" containerID="bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a" Dec 07 10:25:46 crc kubenswrapper[4838]: E1207 10:25:46.908264 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a\": container with ID starting with bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a not found: ID does not exist" containerID="bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.908294 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a"} err="failed to get container status \"bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a\": rpc error: code = NotFound desc = could not find container \"bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a\": container with ID starting with bf2467d43d5f07c4425aa6db4130bbd859d307406ffa39cdab89fbda346a9b1a not found: ID does not exist" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.908315 4838 scope.go:117] "RemoveContainer" containerID="b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1" Dec 07 10:25:46 crc kubenswrapper[4838]: E1207 10:25:46.908857 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1\": container with ID starting with b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1 not found: ID does not exist" containerID="b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.908903 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1"} err="failed to get container status \"b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1\": rpc error: code = NotFound desc = could not find container \"b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1\": container with ID starting with b622b0410e2b6c940b2188696d9c95504fca0e449c599d6d43ec828a17b2e2a1 not found: ID does not exist" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.908933 4838 scope.go:117] "RemoveContainer" containerID="2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793" Dec 07 10:25:46 crc kubenswrapper[4838]: E1207 10:25:46.909440 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793\": container with ID starting with 2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793 not found: ID does not exist" containerID="2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793" Dec 07 10:25:46 crc kubenswrapper[4838]: I1207 10:25:46.909466 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793"} err="failed to get container status \"2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793\": rpc error: code = NotFound desc = could not find container \"2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793\": container with ID starting with 2a580092e72004760239785267567bd7c6b6191a887049c83cf173dce994d793 not found: ID does not exist" Dec 07 10:25:47 crc kubenswrapper[4838]: I1207 10:25:47.628709 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" path="/var/lib/kubelet/pods/4cdbb8bb-a898-4674-8371-b85ad1743a1d/volumes" Dec 07 10:25:54 crc kubenswrapper[4838]: I1207 10:25:54.493654 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:25:54 crc kubenswrapper[4838]: I1207 10:25:54.494206 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:26:24 crc kubenswrapper[4838]: I1207 10:26:24.493691 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:26:24 crc kubenswrapper[4838]: I1207 10:26:24.494630 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:26:24 crc kubenswrapper[4838]: I1207 10:26:24.494719 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 10:26:24 crc kubenswrapper[4838]: I1207 10:26:24.496321 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11045922bb917c10a0c80177be4aee6408cbd108c65ad5b2f6027c8530522657"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 10:26:24 crc kubenswrapper[4838]: I1207 10:26:24.496463 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://11045922bb917c10a0c80177be4aee6408cbd108c65ad5b2f6027c8530522657" gracePeriod=600 Dec 07 10:26:25 crc kubenswrapper[4838]: I1207 10:26:25.163205 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="11045922bb917c10a0c80177be4aee6408cbd108c65ad5b2f6027c8530522657" exitCode=0 Dec 07 10:26:25 crc kubenswrapper[4838]: I1207 10:26:25.163303 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"11045922bb917c10a0c80177be4aee6408cbd108c65ad5b2f6027c8530522657"} Dec 07 10:26:25 crc kubenswrapper[4838]: I1207 10:26:25.163717 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b"} Dec 07 10:26:25 crc kubenswrapper[4838]: I1207 10:26:25.163744 4838 scope.go:117] "RemoveContainer" containerID="0a009d4a7b0186521c3bfbdee26e58f4d0725debf6ec0f69c9926cd2d7c4c0b9" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.223518 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5vfq5"] Dec 07 10:26:55 crc kubenswrapper[4838]: E1207 10:26:55.225025 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="registry-server" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.225052 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="registry-server" Dec 07 10:26:55 crc kubenswrapper[4838]: E1207 10:26:55.225070 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="extract-content" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.225082 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="extract-content" Dec 07 10:26:55 crc kubenswrapper[4838]: E1207 10:26:55.225129 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="extract-utilities" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.225142 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="extract-utilities" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.225487 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cdbb8bb-a898-4674-8371-b85ad1743a1d" containerName="registry-server" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.230622 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.241092 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5vfq5"] Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.396033 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-utilities\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.396449 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rn7\" (UniqueName: \"kubernetes.io/projected/ce746930-07d7-4f29-96be-79c73a7ebeb8-kube-api-access-88rn7\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.396470 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-catalog-content\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.497899 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rn7\" (UniqueName: \"kubernetes.io/projected/ce746930-07d7-4f29-96be-79c73a7ebeb8-kube-api-access-88rn7\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.498179 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-catalog-content\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.498354 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-utilities\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.498707 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-catalog-content\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.498939 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-utilities\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.516630 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rn7\" (UniqueName: \"kubernetes.io/projected/ce746930-07d7-4f29-96be-79c73a7ebeb8-kube-api-access-88rn7\") pod \"community-operators-5vfq5\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:55 crc kubenswrapper[4838]: I1207 10:26:55.571161 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:26:56 crc kubenswrapper[4838]: I1207 10:26:56.008541 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5vfq5"] Dec 07 10:26:56 crc kubenswrapper[4838]: I1207 10:26:56.498181 4838 generic.go:334] "Generic (PLEG): container finished" podID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerID="9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074" exitCode=0 Dec 07 10:26:56 crc kubenswrapper[4838]: I1207 10:26:56.498262 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vfq5" event={"ID":"ce746930-07d7-4f29-96be-79c73a7ebeb8","Type":"ContainerDied","Data":"9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074"} Dec 07 10:26:56 crc kubenswrapper[4838]: I1207 10:26:56.498667 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vfq5" event={"ID":"ce746930-07d7-4f29-96be-79c73a7ebeb8","Type":"ContainerStarted","Data":"6906d9e49ddcfd7d7f22f443e1a9560f3ed811c5e12664934c39da9127f81a0a"} Dec 07 10:26:57 crc kubenswrapper[4838]: I1207 10:26:57.510719 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vfq5" event={"ID":"ce746930-07d7-4f29-96be-79c73a7ebeb8","Type":"ContainerStarted","Data":"74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d"} Dec 07 10:26:58 crc kubenswrapper[4838]: I1207 10:26:58.530596 4838 generic.go:334] "Generic (PLEG): container finished" podID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerID="74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d" exitCode=0 Dec 07 10:26:58 crc kubenswrapper[4838]: I1207 10:26:58.530671 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vfq5" event={"ID":"ce746930-07d7-4f29-96be-79c73a7ebeb8","Type":"ContainerDied","Data":"74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d"} Dec 07 10:26:59 crc kubenswrapper[4838]: I1207 10:26:59.540956 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vfq5" event={"ID":"ce746930-07d7-4f29-96be-79c73a7ebeb8","Type":"ContainerStarted","Data":"0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d"} Dec 07 10:26:59 crc kubenswrapper[4838]: I1207 10:26:59.571540 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5vfq5" podStartSLOduration=2.092073778 podStartE2EDuration="4.571517841s" podCreationTimestamp="2025-12-07 10:26:55 +0000 UTC" firstStartedPulling="2025-12-07 10:26:56.503680311 +0000 UTC m=+4833.210999338" lastFinishedPulling="2025-12-07 10:26:58.983124374 +0000 UTC m=+4835.690443401" observedRunningTime="2025-12-07 10:26:59.564154834 +0000 UTC m=+4836.271473851" watchObservedRunningTime="2025-12-07 10:26:59.571517841 +0000 UTC m=+4836.278836868" Dec 07 10:27:05 crc kubenswrapper[4838]: I1207 10:27:05.572053 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:27:05 crc kubenswrapper[4838]: I1207 10:27:05.572603 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:27:05 crc kubenswrapper[4838]: I1207 10:27:05.670437 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:27:05 crc kubenswrapper[4838]: I1207 10:27:05.746377 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:27:05 crc kubenswrapper[4838]: I1207 10:27:05.919328 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5vfq5"] Dec 07 10:27:07 crc kubenswrapper[4838]: I1207 10:27:07.615324 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5vfq5" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="registry-server" containerID="cri-o://0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d" gracePeriod=2 Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.097171 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.204783 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88rn7\" (UniqueName: \"kubernetes.io/projected/ce746930-07d7-4f29-96be-79c73a7ebeb8-kube-api-access-88rn7\") pod \"ce746930-07d7-4f29-96be-79c73a7ebeb8\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.205160 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-catalog-content\") pod \"ce746930-07d7-4f29-96be-79c73a7ebeb8\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.205314 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-utilities\") pod \"ce746930-07d7-4f29-96be-79c73a7ebeb8\" (UID: \"ce746930-07d7-4f29-96be-79c73a7ebeb8\") " Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.206573 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-utilities" (OuterVolumeSpecName: "utilities") pod "ce746930-07d7-4f29-96be-79c73a7ebeb8" (UID: "ce746930-07d7-4f29-96be-79c73a7ebeb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.224086 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce746930-07d7-4f29-96be-79c73a7ebeb8-kube-api-access-88rn7" (OuterVolumeSpecName: "kube-api-access-88rn7") pod "ce746930-07d7-4f29-96be-79c73a7ebeb8" (UID: "ce746930-07d7-4f29-96be-79c73a7ebeb8"). InnerVolumeSpecName "kube-api-access-88rn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.282446 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce746930-07d7-4f29-96be-79c73a7ebeb8" (UID: "ce746930-07d7-4f29-96be-79c73a7ebeb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.308376 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.308437 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88rn7\" (UniqueName: \"kubernetes.io/projected/ce746930-07d7-4f29-96be-79c73a7ebeb8-kube-api-access-88rn7\") on node \"crc\" DevicePath \"\"" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.308459 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce746930-07d7-4f29-96be-79c73a7ebeb8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.655970 4838 generic.go:334] "Generic (PLEG): container finished" podID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerID="0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d" exitCode=0 Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.656023 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vfq5" event={"ID":"ce746930-07d7-4f29-96be-79c73a7ebeb8","Type":"ContainerDied","Data":"0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d"} Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.656056 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5vfq5" event={"ID":"ce746930-07d7-4f29-96be-79c73a7ebeb8","Type":"ContainerDied","Data":"6906d9e49ddcfd7d7f22f443e1a9560f3ed811c5e12664934c39da9127f81a0a"} Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.656077 4838 scope.go:117] "RemoveContainer" containerID="0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.657694 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5vfq5" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.688245 4838 scope.go:117] "RemoveContainer" containerID="74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.716788 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5vfq5"] Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.726410 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5vfq5"] Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.736117 4838 scope.go:117] "RemoveContainer" containerID="9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.772108 4838 scope.go:117] "RemoveContainer" containerID="0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d" Dec 07 10:27:08 crc kubenswrapper[4838]: E1207 10:27:08.772759 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d\": container with ID starting with 0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d not found: ID does not exist" containerID="0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.772806 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d"} err="failed to get container status \"0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d\": rpc error: code = NotFound desc = could not find container \"0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d\": container with ID starting with 0e8bdbc98a8f9bc530f8da94c20dbb7495506f082565127583ce02ce6d16fa2d not found: ID does not exist" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.772925 4838 scope.go:117] "RemoveContainer" containerID="74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d" Dec 07 10:27:08 crc kubenswrapper[4838]: E1207 10:27:08.773430 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d\": container with ID starting with 74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d not found: ID does not exist" containerID="74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.773521 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d"} err="failed to get container status \"74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d\": rpc error: code = NotFound desc = could not find container \"74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d\": container with ID starting with 74c0ec73f0e92e3118b5b9c36eb1c357a4916f6db3db3d1aa781ffb7d857ae7d not found: ID does not exist" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.773572 4838 scope.go:117] "RemoveContainer" containerID="9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074" Dec 07 10:27:08 crc kubenswrapper[4838]: E1207 10:27:08.774073 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074\": container with ID starting with 9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074 not found: ID does not exist" containerID="9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074" Dec 07 10:27:08 crc kubenswrapper[4838]: I1207 10:27:08.774145 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074"} err="failed to get container status \"9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074\": rpc error: code = NotFound desc = could not find container \"9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074\": container with ID starting with 9e1e800cf504bd5eb0b7208a334f4dbd1e3d96f39489315c151cb8ab9a051074 not found: ID does not exist" Dec 07 10:27:09 crc kubenswrapper[4838]: I1207 10:27:09.638629 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" path="/var/lib/kubelet/pods/ce746930-07d7-4f29-96be-79c73a7ebeb8/volumes" Dec 07 10:28:24 crc kubenswrapper[4838]: I1207 10:28:24.493350 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:28:24 crc kubenswrapper[4838]: I1207 10:28:24.494881 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.689273 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ttxz8/must-gather-wttxv"] Dec 07 10:28:27 crc kubenswrapper[4838]: E1207 10:28:27.690306 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="registry-server" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.690324 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="registry-server" Dec 07 10:28:27 crc kubenswrapper[4838]: E1207 10:28:27.690344 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="extract-utilities" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.690353 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="extract-utilities" Dec 07 10:28:27 crc kubenswrapper[4838]: E1207 10:28:27.690371 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="extract-content" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.690379 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="extract-content" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.690620 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce746930-07d7-4f29-96be-79c73a7ebeb8" containerName="registry-server" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.691849 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.696924 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ttxz8"/"kube-root-ca.crt" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.697312 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ttxz8"/"openshift-service-ca.crt" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.716178 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ttxz8/must-gather-wttxv"] Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.801720 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-must-gather-output\") pod \"must-gather-wttxv\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.801931 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shll6\" (UniqueName: \"kubernetes.io/projected/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-kube-api-access-shll6\") pod \"must-gather-wttxv\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.903979 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-must-gather-output\") pod \"must-gather-wttxv\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.904389 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shll6\" (UniqueName: \"kubernetes.io/projected/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-kube-api-access-shll6\") pod \"must-gather-wttxv\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.904469 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-must-gather-output\") pod \"must-gather-wttxv\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:27 crc kubenswrapper[4838]: I1207 10:28:27.930795 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shll6\" (UniqueName: \"kubernetes.io/projected/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-kube-api-access-shll6\") pod \"must-gather-wttxv\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:28 crc kubenswrapper[4838]: I1207 10:28:28.014749 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:28:28 crc kubenswrapper[4838]: I1207 10:28:28.496601 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ttxz8/must-gather-wttxv"] Dec 07 10:28:28 crc kubenswrapper[4838]: W1207 10:28:28.502487 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3766e4fb_8a36_431a_9cfe_8984b5d60f3b.slice/crio-4138f137b1a530b3460809b61ba9f426402fd32544dec4a2f0791f473304bb66 WatchSource:0}: Error finding container 4138f137b1a530b3460809b61ba9f426402fd32544dec4a2f0791f473304bb66: Status 404 returned error can't find the container with id 4138f137b1a530b3460809b61ba9f426402fd32544dec4a2f0791f473304bb66 Dec 07 10:28:28 crc kubenswrapper[4838]: I1207 10:28:28.675433 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/must-gather-wttxv" event={"ID":"3766e4fb-8a36-431a-9cfe-8984b5d60f3b","Type":"ContainerStarted","Data":"4138f137b1a530b3460809b61ba9f426402fd32544dec4a2f0791f473304bb66"} Dec 07 10:28:29 crc kubenswrapper[4838]: I1207 10:28:29.688946 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/must-gather-wttxv" event={"ID":"3766e4fb-8a36-431a-9cfe-8984b5d60f3b","Type":"ContainerStarted","Data":"357dbe7a0d2abba6795b387e7e1a7d04cbcadf0b9c48d0627a326af3ea99460a"} Dec 07 10:28:29 crc kubenswrapper[4838]: I1207 10:28:29.689665 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/must-gather-wttxv" event={"ID":"3766e4fb-8a36-431a-9cfe-8984b5d60f3b","Type":"ContainerStarted","Data":"73208fc40e1d34e46ee779e0e6d0364ace68556c65f4cfb1d2db33ce85a7c12e"} Dec 07 10:28:29 crc kubenswrapper[4838]: I1207 10:28:29.713096 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ttxz8/must-gather-wttxv" podStartSLOduration=2.713069145 podStartE2EDuration="2.713069145s" podCreationTimestamp="2025-12-07 10:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:28:29.707313193 +0000 UTC m=+4926.414632210" watchObservedRunningTime="2025-12-07 10:28:29.713069145 +0000 UTC m=+4926.420388182" Dec 07 10:28:31 crc kubenswrapper[4838]: E1207 10:28:31.992253 4838 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.73:36930->38.102.83.73:33371: write tcp 38.102.83.73:36930->38.102.83.73:33371: write: broken pipe Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.433557 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-m9sx6"] Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.436549 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.438583 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ttxz8"/"default-dockercfg-7mj64" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.519593 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3a5675a-dfde-4875-9e1b-d8e4677a5539-host\") pod \"crc-debug-m9sx6\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.519643 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndkbs\" (UniqueName: \"kubernetes.io/projected/d3a5675a-dfde-4875-9e1b-d8e4677a5539-kube-api-access-ndkbs\") pod \"crc-debug-m9sx6\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.621312 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3a5675a-dfde-4875-9e1b-d8e4677a5539-host\") pod \"crc-debug-m9sx6\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.621354 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndkbs\" (UniqueName: \"kubernetes.io/projected/d3a5675a-dfde-4875-9e1b-d8e4677a5539-kube-api-access-ndkbs\") pod \"crc-debug-m9sx6\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.621441 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3a5675a-dfde-4875-9e1b-d8e4677a5539-host\") pod \"crc-debug-m9sx6\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.640981 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndkbs\" (UniqueName: \"kubernetes.io/projected/d3a5675a-dfde-4875-9e1b-d8e4677a5539-kube-api-access-ndkbs\") pod \"crc-debug-m9sx6\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:33 crc kubenswrapper[4838]: I1207 10:28:33.752976 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:28:34 crc kubenswrapper[4838]: I1207 10:28:34.730477 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" event={"ID":"d3a5675a-dfde-4875-9e1b-d8e4677a5539","Type":"ContainerStarted","Data":"896037a86a958ec1b5cbf5d76bf39b8c7e8cd2dd4d85802ac4cca5b93b01f575"} Dec 07 10:28:34 crc kubenswrapper[4838]: I1207 10:28:34.731130 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" event={"ID":"d3a5675a-dfde-4875-9e1b-d8e4677a5539","Type":"ContainerStarted","Data":"067959818ba4490a781ebbb054a0c8f4d25f71e43c9aa3e3a145e6e27b609ced"} Dec 07 10:28:34 crc kubenswrapper[4838]: I1207 10:28:34.753544 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" podStartSLOduration=1.75352405 podStartE2EDuration="1.75352405s" podCreationTimestamp="2025-12-07 10:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 10:28:34.751233845 +0000 UTC m=+4931.458552882" watchObservedRunningTime="2025-12-07 10:28:34.75352405 +0000 UTC m=+4931.460843087" Dec 07 10:28:54 crc kubenswrapper[4838]: I1207 10:28:54.493423 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:28:54 crc kubenswrapper[4838]: I1207 10:28:54.493957 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:29:06 crc kubenswrapper[4838]: I1207 10:29:06.066409 4838 generic.go:334] "Generic (PLEG): container finished" podID="d3a5675a-dfde-4875-9e1b-d8e4677a5539" containerID="896037a86a958ec1b5cbf5d76bf39b8c7e8cd2dd4d85802ac4cca5b93b01f575" exitCode=0 Dec 07 10:29:06 crc kubenswrapper[4838]: I1207 10:29:06.066506 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" event={"ID":"d3a5675a-dfde-4875-9e1b-d8e4677a5539","Type":"ContainerDied","Data":"896037a86a958ec1b5cbf5d76bf39b8c7e8cd2dd4d85802ac4cca5b93b01f575"} Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.172586 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.201063 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-m9sx6"] Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.209197 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-m9sx6"] Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.297190 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3a5675a-dfde-4875-9e1b-d8e4677a5539-host\") pod \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.297271 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3a5675a-dfde-4875-9e1b-d8e4677a5539-host" (OuterVolumeSpecName: "host") pod "d3a5675a-dfde-4875-9e1b-d8e4677a5539" (UID: "d3a5675a-dfde-4875-9e1b-d8e4677a5539"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.298104 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndkbs\" (UniqueName: \"kubernetes.io/projected/d3a5675a-dfde-4875-9e1b-d8e4677a5539-kube-api-access-ndkbs\") pod \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\" (UID: \"d3a5675a-dfde-4875-9e1b-d8e4677a5539\") " Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.298846 4838 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3a5675a-dfde-4875-9e1b-d8e4677a5539-host\") on node \"crc\" DevicePath \"\"" Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.304127 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a5675a-dfde-4875-9e1b-d8e4677a5539-kube-api-access-ndkbs" (OuterVolumeSpecName: "kube-api-access-ndkbs") pod "d3a5675a-dfde-4875-9e1b-d8e4677a5539" (UID: "d3a5675a-dfde-4875-9e1b-d8e4677a5539"). InnerVolumeSpecName "kube-api-access-ndkbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.400965 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndkbs\" (UniqueName: \"kubernetes.io/projected/d3a5675a-dfde-4875-9e1b-d8e4677a5539-kube-api-access-ndkbs\") on node \"crc\" DevicePath \"\"" Dec 07 10:29:07 crc kubenswrapper[4838]: I1207 10:29:07.624491 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a5675a-dfde-4875-9e1b-d8e4677a5539" path="/var/lib/kubelet/pods/d3a5675a-dfde-4875-9e1b-d8e4677a5539/volumes" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.084275 4838 scope.go:117] "RemoveContainer" containerID="896037a86a958ec1b5cbf5d76bf39b8c7e8cd2dd4d85802ac4cca5b93b01f575" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.084305 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-m9sx6" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.425554 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-nk2j6"] Dec 07 10:29:08 crc kubenswrapper[4838]: E1207 10:29:08.426260 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a5675a-dfde-4875-9e1b-d8e4677a5539" containerName="container-00" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.426273 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a5675a-dfde-4875-9e1b-d8e4677a5539" containerName="container-00" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.426435 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a5675a-dfde-4875-9e1b-d8e4677a5539" containerName="container-00" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.427058 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.428721 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ttxz8"/"default-dockercfg-7mj64" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.524051 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ksqh\" (UniqueName: \"kubernetes.io/projected/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-kube-api-access-7ksqh\") pod \"crc-debug-nk2j6\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.524132 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-host\") pod \"crc-debug-nk2j6\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.626113 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ksqh\" (UniqueName: \"kubernetes.io/projected/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-kube-api-access-7ksqh\") pod \"crc-debug-nk2j6\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.626171 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-host\") pod \"crc-debug-nk2j6\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:08 crc kubenswrapper[4838]: I1207 10:29:08.626370 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-host\") pod \"crc-debug-nk2j6\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:09 crc kubenswrapper[4838]: I1207 10:29:09.016859 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ksqh\" (UniqueName: \"kubernetes.io/projected/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-kube-api-access-7ksqh\") pod \"crc-debug-nk2j6\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:09 crc kubenswrapper[4838]: I1207 10:29:09.042765 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:10 crc kubenswrapper[4838]: I1207 10:29:10.103451 4838 generic.go:334] "Generic (PLEG): container finished" podID="8c8767b8-ba30-49cf-8ae4-d54b1db91f09" containerID="e1083b2b1a58cdbbd5cc89dee9ebc35dfce988e49692e26c98ee7660331ea1bf" exitCode=0 Dec 07 10:29:10 crc kubenswrapper[4838]: I1207 10:29:10.103552 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" event={"ID":"8c8767b8-ba30-49cf-8ae4-d54b1db91f09","Type":"ContainerDied","Data":"e1083b2b1a58cdbbd5cc89dee9ebc35dfce988e49692e26c98ee7660331ea1bf"} Dec 07 10:29:10 crc kubenswrapper[4838]: I1207 10:29:10.103951 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" event={"ID":"8c8767b8-ba30-49cf-8ae4-d54b1db91f09","Type":"ContainerStarted","Data":"12a5a894b250353562132acb5cfd4ea2e9ccf75d2d7e2f0c167af296c398bd6e"} Dec 07 10:29:10 crc kubenswrapper[4838]: I1207 10:29:10.425538 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-nk2j6"] Dec 07 10:29:10 crc kubenswrapper[4838]: I1207 10:29:10.441495 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-nk2j6"] Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.207152 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.280697 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ksqh\" (UniqueName: \"kubernetes.io/projected/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-kube-api-access-7ksqh\") pod \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.280838 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-host\") pod \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\" (UID: \"8c8767b8-ba30-49cf-8ae4-d54b1db91f09\") " Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.281219 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-host" (OuterVolumeSpecName: "host") pod "8c8767b8-ba30-49cf-8ae4-d54b1db91f09" (UID: "8c8767b8-ba30-49cf-8ae4-d54b1db91f09"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.281913 4838 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-host\") on node \"crc\" DevicePath \"\"" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.291990 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-kube-api-access-7ksqh" (OuterVolumeSpecName: "kube-api-access-7ksqh") pod "8c8767b8-ba30-49cf-8ae4-d54b1db91f09" (UID: "8c8767b8-ba30-49cf-8ae4-d54b1db91f09"). InnerVolumeSpecName "kube-api-access-7ksqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.383427 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ksqh\" (UniqueName: \"kubernetes.io/projected/8c8767b8-ba30-49cf-8ae4-d54b1db91f09-kube-api-access-7ksqh\") on node \"crc\" DevicePath \"\"" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.631918 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c8767b8-ba30-49cf-8ae4-d54b1db91f09" path="/var/lib/kubelet/pods/8c8767b8-ba30-49cf-8ae4-d54b1db91f09/volumes" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.958387 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-kxvh4"] Dec 07 10:29:11 crc kubenswrapper[4838]: E1207 10:29:11.958877 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8767b8-ba30-49cf-8ae4-d54b1db91f09" containerName="container-00" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.958893 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8767b8-ba30-49cf-8ae4-d54b1db91f09" containerName="container-00" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.959132 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c8767b8-ba30-49cf-8ae4-d54b1db91f09" containerName="container-00" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.959940 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.996172 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/442d3f4c-0795-47c7-9438-4d6048f68616-host\") pod \"crc-debug-kxvh4\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:11 crc kubenswrapper[4838]: I1207 10:29:11.996342 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c75t\" (UniqueName: \"kubernetes.io/projected/442d3f4c-0795-47c7-9438-4d6048f68616-kube-api-access-8c75t\") pod \"crc-debug-kxvh4\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:12 crc kubenswrapper[4838]: I1207 10:29:12.097855 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c75t\" (UniqueName: \"kubernetes.io/projected/442d3f4c-0795-47c7-9438-4d6048f68616-kube-api-access-8c75t\") pod \"crc-debug-kxvh4\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:12 crc kubenswrapper[4838]: I1207 10:29:12.097933 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/442d3f4c-0795-47c7-9438-4d6048f68616-host\") pod \"crc-debug-kxvh4\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:12 crc kubenswrapper[4838]: I1207 10:29:12.098322 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/442d3f4c-0795-47c7-9438-4d6048f68616-host\") pod \"crc-debug-kxvh4\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:12 crc kubenswrapper[4838]: I1207 10:29:12.114438 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c75t\" (UniqueName: \"kubernetes.io/projected/442d3f4c-0795-47c7-9438-4d6048f68616-kube-api-access-8c75t\") pod \"crc-debug-kxvh4\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:12 crc kubenswrapper[4838]: I1207 10:29:12.118777 4838 scope.go:117] "RemoveContainer" containerID="e1083b2b1a58cdbbd5cc89dee9ebc35dfce988e49692e26c98ee7660331ea1bf" Dec 07 10:29:12 crc kubenswrapper[4838]: I1207 10:29:12.118801 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-nk2j6" Dec 07 10:29:12 crc kubenswrapper[4838]: I1207 10:29:12.287639 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:12 crc kubenswrapper[4838]: W1207 10:29:12.324790 4838 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod442d3f4c_0795_47c7_9438_4d6048f68616.slice/crio-21003d14d613b360535806edea758112eee8f6cf4047c489f54e5d2ef804c178 WatchSource:0}: Error finding container 21003d14d613b360535806edea758112eee8f6cf4047c489f54e5d2ef804c178: Status 404 returned error can't find the container with id 21003d14d613b360535806edea758112eee8f6cf4047c489f54e5d2ef804c178 Dec 07 10:29:13 crc kubenswrapper[4838]: I1207 10:29:13.128661 4838 generic.go:334] "Generic (PLEG): container finished" podID="442d3f4c-0795-47c7-9438-4d6048f68616" containerID="62c75fb6e373530baa8e7ca1d166a2a937a2c740bb218eec8db1f87a081787fe" exitCode=0 Dec 07 10:29:13 crc kubenswrapper[4838]: I1207 10:29:13.128771 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" event={"ID":"442d3f4c-0795-47c7-9438-4d6048f68616","Type":"ContainerDied","Data":"62c75fb6e373530baa8e7ca1d166a2a937a2c740bb218eec8db1f87a081787fe"} Dec 07 10:29:13 crc kubenswrapper[4838]: I1207 10:29:13.129008 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" event={"ID":"442d3f4c-0795-47c7-9438-4d6048f68616","Type":"ContainerStarted","Data":"21003d14d613b360535806edea758112eee8f6cf4047c489f54e5d2ef804c178"} Dec 07 10:29:13 crc kubenswrapper[4838]: I1207 10:29:13.167139 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-kxvh4"] Dec 07 10:29:13 crc kubenswrapper[4838]: I1207 10:29:13.179197 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ttxz8/crc-debug-kxvh4"] Dec 07 10:29:14 crc kubenswrapper[4838]: I1207 10:29:14.228575 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:14 crc kubenswrapper[4838]: I1207 10:29:14.239981 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/442d3f4c-0795-47c7-9438-4d6048f68616-host\") pod \"442d3f4c-0795-47c7-9438-4d6048f68616\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " Dec 07 10:29:14 crc kubenswrapper[4838]: I1207 10:29:14.240143 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c75t\" (UniqueName: \"kubernetes.io/projected/442d3f4c-0795-47c7-9438-4d6048f68616-kube-api-access-8c75t\") pod \"442d3f4c-0795-47c7-9438-4d6048f68616\" (UID: \"442d3f4c-0795-47c7-9438-4d6048f68616\") " Dec 07 10:29:14 crc kubenswrapper[4838]: I1207 10:29:14.240348 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/442d3f4c-0795-47c7-9438-4d6048f68616-host" (OuterVolumeSpecName: "host") pod "442d3f4c-0795-47c7-9438-4d6048f68616" (UID: "442d3f4c-0795-47c7-9438-4d6048f68616"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 10:29:14 crc kubenswrapper[4838]: I1207 10:29:14.240800 4838 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/442d3f4c-0795-47c7-9438-4d6048f68616-host\") on node \"crc\" DevicePath \"\"" Dec 07 10:29:14 crc kubenswrapper[4838]: I1207 10:29:14.247097 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442d3f4c-0795-47c7-9438-4d6048f68616-kube-api-access-8c75t" (OuterVolumeSpecName: "kube-api-access-8c75t") pod "442d3f4c-0795-47c7-9438-4d6048f68616" (UID: "442d3f4c-0795-47c7-9438-4d6048f68616"). InnerVolumeSpecName "kube-api-access-8c75t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:29:14 crc kubenswrapper[4838]: I1207 10:29:14.342297 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c75t\" (UniqueName: \"kubernetes.io/projected/442d3f4c-0795-47c7-9438-4d6048f68616-kube-api-access-8c75t\") on node \"crc\" DevicePath \"\"" Dec 07 10:29:15 crc kubenswrapper[4838]: I1207 10:29:15.145827 4838 scope.go:117] "RemoveContainer" containerID="62c75fb6e373530baa8e7ca1d166a2a937a2c740bb218eec8db1f87a081787fe" Dec 07 10:29:15 crc kubenswrapper[4838]: I1207 10:29:15.146010 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/crc-debug-kxvh4" Dec 07 10:29:15 crc kubenswrapper[4838]: I1207 10:29:15.624209 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="442d3f4c-0795-47c7-9438-4d6048f68616" path="/var/lib/kubelet/pods/442d3f4c-0795-47c7-9438-4d6048f68616/volumes" Dec 07 10:29:24 crc kubenswrapper[4838]: I1207 10:29:24.493268 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:29:24 crc kubenswrapper[4838]: I1207 10:29:24.493701 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:29:24 crc kubenswrapper[4838]: I1207 10:29:24.493756 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 10:29:24 crc kubenswrapper[4838]: I1207 10:29:24.494637 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 10:29:24 crc kubenswrapper[4838]: I1207 10:29:24.494682 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" gracePeriod=600 Dec 07 10:29:24 crc kubenswrapper[4838]: E1207 10:29:24.633374 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:29:25 crc kubenswrapper[4838]: I1207 10:29:25.234563 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" exitCode=0 Dec 07 10:29:25 crc kubenswrapper[4838]: I1207 10:29:25.234723 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b"} Dec 07 10:29:25 crc kubenswrapper[4838]: I1207 10:29:25.234924 4838 scope.go:117] "RemoveContainer" containerID="11045922bb917c10a0c80177be4aee6408cbd108c65ad5b2f6027c8530522657" Dec 07 10:29:25 crc kubenswrapper[4838]: I1207 10:29:25.235644 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:29:25 crc kubenswrapper[4838]: E1207 10:29:25.235949 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:29:39 crc kubenswrapper[4838]: I1207 10:29:39.615003 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:29:39 crc kubenswrapper[4838]: E1207 10:29:39.615774 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:29:52 crc kubenswrapper[4838]: I1207 10:29:52.614533 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:29:52 crc kubenswrapper[4838]: E1207 10:29:52.615229 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.162262 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2"] Dec 07 10:30:00 crc kubenswrapper[4838]: E1207 10:30:00.163260 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442d3f4c-0795-47c7-9438-4d6048f68616" containerName="container-00" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.163279 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="442d3f4c-0795-47c7-9438-4d6048f68616" containerName="container-00" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.163534 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="442d3f4c-0795-47c7-9438-4d6048f68616" containerName="container-00" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.166944 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.169905 4838 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.182562 4838 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.191405 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2"] Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.196386 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96023c1b-8bce-4ab1-b5f2-458ea9113367-config-volume\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.196467 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96023c1b-8bce-4ab1-b5f2-458ea9113367-secret-volume\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.196542 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxtzc\" (UniqueName: \"kubernetes.io/projected/96023c1b-8bce-4ab1-b5f2-458ea9113367-kube-api-access-cxtzc\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.298190 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96023c1b-8bce-4ab1-b5f2-458ea9113367-config-volume\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.298465 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96023c1b-8bce-4ab1-b5f2-458ea9113367-secret-volume\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.298598 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxtzc\" (UniqueName: \"kubernetes.io/projected/96023c1b-8bce-4ab1-b5f2-458ea9113367-kube-api-access-cxtzc\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.299210 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96023c1b-8bce-4ab1-b5f2-458ea9113367-config-volume\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.310427 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96023c1b-8bce-4ab1-b5f2-458ea9113367-secret-volume\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.322267 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxtzc\" (UniqueName: \"kubernetes.io/projected/96023c1b-8bce-4ab1-b5f2-458ea9113367-kube-api-access-cxtzc\") pod \"collect-profiles-29418390-65hr2\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.490598 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:00 crc kubenswrapper[4838]: I1207 10:30:00.960774 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2"] Dec 07 10:30:01 crc kubenswrapper[4838]: I1207 10:30:01.603333 4838 generic.go:334] "Generic (PLEG): container finished" podID="96023c1b-8bce-4ab1-b5f2-458ea9113367" containerID="fad290b0f13ca941f982c0ab674f78ec7f93146b10089779377c75ae46481129" exitCode=0 Dec 07 10:30:01 crc kubenswrapper[4838]: I1207 10:30:01.603548 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" event={"ID":"96023c1b-8bce-4ab1-b5f2-458ea9113367","Type":"ContainerDied","Data":"fad290b0f13ca941f982c0ab674f78ec7f93146b10089779377c75ae46481129"} Dec 07 10:30:01 crc kubenswrapper[4838]: I1207 10:30:01.603808 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" event={"ID":"96023c1b-8bce-4ab1-b5f2-458ea9113367","Type":"ContainerStarted","Data":"649cb49ddb04ea815435c1a9f12fd9afedf80fb7588991a86ae73432be5de885"} Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.114947 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.262307 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96023c1b-8bce-4ab1-b5f2-458ea9113367-config-volume\") pod \"96023c1b-8bce-4ab1-b5f2-458ea9113367\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.262648 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxtzc\" (UniqueName: \"kubernetes.io/projected/96023c1b-8bce-4ab1-b5f2-458ea9113367-kube-api-access-cxtzc\") pod \"96023c1b-8bce-4ab1-b5f2-458ea9113367\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.262721 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96023c1b-8bce-4ab1-b5f2-458ea9113367-secret-volume\") pod \"96023c1b-8bce-4ab1-b5f2-458ea9113367\" (UID: \"96023c1b-8bce-4ab1-b5f2-458ea9113367\") " Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.263898 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96023c1b-8bce-4ab1-b5f2-458ea9113367-config-volume" (OuterVolumeSpecName: "config-volume") pod "96023c1b-8bce-4ab1-b5f2-458ea9113367" (UID: "96023c1b-8bce-4ab1-b5f2-458ea9113367"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.274170 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96023c1b-8bce-4ab1-b5f2-458ea9113367-kube-api-access-cxtzc" (OuterVolumeSpecName: "kube-api-access-cxtzc") pod "96023c1b-8bce-4ab1-b5f2-458ea9113367" (UID: "96023c1b-8bce-4ab1-b5f2-458ea9113367"). InnerVolumeSpecName "kube-api-access-cxtzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.274280 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96023c1b-8bce-4ab1-b5f2-458ea9113367-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "96023c1b-8bce-4ab1-b5f2-458ea9113367" (UID: "96023c1b-8bce-4ab1-b5f2-458ea9113367"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.366386 4838 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96023c1b-8bce-4ab1-b5f2-458ea9113367-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.366448 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxtzc\" (UniqueName: \"kubernetes.io/projected/96023c1b-8bce-4ab1-b5f2-458ea9113367-kube-api-access-cxtzc\") on node \"crc\" DevicePath \"\"" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.366476 4838 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96023c1b-8bce-4ab1-b5f2-458ea9113367-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.627961 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" event={"ID":"96023c1b-8bce-4ab1-b5f2-458ea9113367","Type":"ContainerDied","Data":"649cb49ddb04ea815435c1a9f12fd9afedf80fb7588991a86ae73432be5de885"} Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.628008 4838 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="649cb49ddb04ea815435c1a9f12fd9afedf80fb7588991a86ae73432be5de885" Dec 07 10:30:03 crc kubenswrapper[4838]: I1207 10:30:03.628073 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418390-65hr2" Dec 07 10:30:04 crc kubenswrapper[4838]: I1207 10:30:04.225176 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg"] Dec 07 10:30:04 crc kubenswrapper[4838]: I1207 10:30:04.240726 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418345-m6chg"] Dec 07 10:30:04 crc kubenswrapper[4838]: I1207 10:30:04.615574 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:30:04 crc kubenswrapper[4838]: E1207 10:30:04.616670 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:30:05 crc kubenswrapper[4838]: I1207 10:30:05.631589 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aadda5c8-2133-474c-a880-4a3c519bf748" path="/var/lib/kubelet/pods/aadda5c8-2133-474c-a880-4a3c519bf748/volumes" Dec 07 10:30:17 crc kubenswrapper[4838]: I1207 10:30:17.614552 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:30:17 crc kubenswrapper[4838]: E1207 10:30:17.615399 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:30:30 crc kubenswrapper[4838]: I1207 10:30:30.615332 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:30:30 crc kubenswrapper[4838]: E1207 10:30:30.616299 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.123086 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59c4fb88bd-m7xbq_279b7609-3731-4b2d-a74f-64b591e53925/barbican-api/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.194667 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59c4fb88bd-m7xbq_279b7609-3731-4b2d-a74f-64b591e53925/barbican-api-log/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.337826 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d7849ff8d-sjw2m_2d766abd-30cf-45e4-9672-1f999276a187/barbican-keystone-listener/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.405585 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d7849ff8d-sjw2m_2d766abd-30cf-45e4-9672-1f999276a187/barbican-keystone-listener-log/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.470714 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-698f6d859c-2w95m_86c6d276-ba63-41ba-aa25-10cb0994f54a/barbican-worker/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.550516 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-698f6d859c-2w95m_86c6d276-ba63-41ba-aa25-10cb0994f54a/barbican-worker-log/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.690341 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-8q5sd_1386eb32-0cef-4287-93d1-8fb5445081f3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.765448 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/ceilometer-central-agent/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.904240 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/ceilometer-notification-agent/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.962056 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/proxy-httpd/0.log" Dec 07 10:30:34 crc kubenswrapper[4838]: I1207 10:30:34.971994 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_84d4963b-0485-4254-a707-d621bc87bf6d/sg-core/0.log" Dec 07 10:30:35 crc kubenswrapper[4838]: I1207 10:30:35.654646 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-5qr8d_d681f081-977d-48c0-ac82-ff3c61647de8/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:35 crc kubenswrapper[4838]: I1207 10:30:35.674542 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7xd56_9e6c35ea-c607-4c61-84d0-7bdf0d62add9/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.230650 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7e943cd3-c74f-4317-951f-5d1696ee7ffd/cinder-api-log/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.263731 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7e943cd3-c74f-4317-951f-5d1696ee7ffd/cinder-api/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.523646 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5d85f239-60d0-45b7-b884-a3b625160cc0/probe/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.606199 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5d85f239-60d0-45b7-b884-a3b625160cc0/cinder-backup/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.651788 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51d41156-ff39-4850-a562-5ba714cbd169/cinder-scheduler/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.811832 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51d41156-ff39-4850-a562-5ba714cbd169/probe/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.953782 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_874c3661-01ec-4c0c-8929-e05fc0fcf66e/cinder-volume/0.log" Dec 07 10:30:36 crc kubenswrapper[4838]: I1207 10:30:36.984475 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_874c3661-01ec-4c0c-8929-e05fc0fcf66e/probe/0.log" Dec 07 10:30:37 crc kubenswrapper[4838]: I1207 10:30:37.260249 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-m5dsk_f5b4b264-d902-41a3-a100-6455a8a7b6ca/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:37 crc kubenswrapper[4838]: I1207 10:30:37.280984 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-mf8lq_687f9d1b-115c-4b7a-bd3c-35666e665ec3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:37 crc kubenswrapper[4838]: I1207 10:30:37.596098 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7944d6f96c-r7hkg_2e347003-e564-4cf3-b006-9b4b3daf4044/init/0.log" Dec 07 10:30:37 crc kubenswrapper[4838]: I1207 10:30:37.806612 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7944d6f96c-r7hkg_2e347003-e564-4cf3-b006-9b4b3daf4044/init/0.log" Dec 07 10:30:37 crc kubenswrapper[4838]: I1207 10:30:37.929397 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7944d6f96c-r7hkg_2e347003-e564-4cf3-b006-9b4b3daf4044/dnsmasq-dns/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.020871 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f8f155e0-a710-40b0-af6b-157bb7e25854/glance-httpd/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.138425 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f8f155e0-a710-40b0-af6b-157bb7e25854/glance-log/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.249261 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_57a396f6-7966-45b4-a4a5-4f91e688eaaa/glance-httpd/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.309190 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_57a396f6-7966-45b4-a4a5-4f91e688eaaa/glance-log/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.533356 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-76f6479d76-n4xcx_3f6fe9f7-9f77-4da5-823b-b650957c5860/horizon/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.684431 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-76f6479d76-n4xcx_3f6fe9f7-9f77-4da5-823b-b650957c5860/horizon-log/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.743002 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-kqh2s_84df8337-c690-4c33-9d82-da53b4de5be5/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:38 crc kubenswrapper[4838]: I1207 10:30:38.918268 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nlmxt_6b76e0a9-db0a-4913-90a7-1306215b82b7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:39 crc kubenswrapper[4838]: I1207 10:30:39.013145 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29418361-mnpmk_5de7aef2-84f7-4398-a557-4041f9031e31/keystone-cron/0.log" Dec 07 10:30:39 crc kubenswrapper[4838]: I1207 10:30:39.160247 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-654fd9b8c-czchm_dc215482-4e46-4b74-83da-82e55658e2d6/keystone-api/0.log" Dec 07 10:30:39 crc kubenswrapper[4838]: I1207 10:30:39.265875 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a10a8e53-204b-48d6-aaef-809d1847f251/kube-state-metrics/0.log" Dec 07 10:30:39 crc kubenswrapper[4838]: I1207 10:30:39.421138 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xsvzt_47331703-36e3-47e2-bbe0-e891130bdbe7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:39 crc kubenswrapper[4838]: I1207 10:30:39.751336 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_79c0afb9-cd0a-4608-87d1-1163282d784a/manila-api-log/0.log" Dec 07 10:30:39 crc kubenswrapper[4838]: I1207 10:30:39.962695 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_79c0afb9-cd0a-4608-87d1-1163282d784a/manila-api/0.log" Dec 07 10:30:39 crc kubenswrapper[4838]: I1207 10:30:39.966358 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_f404f222-0d07-479e-a0d9-0cb1ce143f3e/probe/0.log" Dec 07 10:30:40 crc kubenswrapper[4838]: I1207 10:30:40.011510 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_f404f222-0d07-479e-a0d9-0cb1ce143f3e/manila-scheduler/0.log" Dec 07 10:30:40 crc kubenswrapper[4838]: I1207 10:30:40.244732 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b5e3f331-938e-4ca9-842e-a580af7035d2/manila-share/0.log" Dec 07 10:30:40 crc kubenswrapper[4838]: I1207 10:30:40.246660 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b5e3f331-938e-4ca9-842e-a580af7035d2/probe/0.log" Dec 07 10:30:40 crc kubenswrapper[4838]: I1207 10:30:40.624621 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6zzng_63630234-9950-43db-a3da-a4ea9ba22be3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:40 crc kubenswrapper[4838]: I1207 10:30:40.635965 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7998cf6c8f-xbbtt_562c9fcd-8827-42a8-8753-d570d77656fc/neutron-api/0.log" Dec 07 10:30:40 crc kubenswrapper[4838]: I1207 10:30:40.681488 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7998cf6c8f-xbbtt_562c9fcd-8827-42a8-8753-d570d77656fc/neutron-httpd/0.log" Dec 07 10:30:41 crc kubenswrapper[4838]: I1207 10:30:41.317992 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1d429890-5092-448b-b386-36493df22443/nova-api-log/0.log" Dec 07 10:30:41 crc kubenswrapper[4838]: I1207 10:30:41.376223 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_485d8af5-c6b8-4d04-be66-a3a421f86d49/nova-cell0-conductor-conductor/0.log" Dec 07 10:30:41 crc kubenswrapper[4838]: I1207 10:30:41.762757 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_8cdfa44f-0d15-4338-8a82-82b624b9ba6d/nova-cell1-conductor-conductor/0.log" Dec 07 10:30:41 crc kubenswrapper[4838]: I1207 10:30:41.893004 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_cb3ea478-6a63-4e03-b9c1-890fca36b984/nova-cell1-novncproxy-novncproxy/0.log" Dec 07 10:30:41 crc kubenswrapper[4838]: I1207 10:30:41.934535 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1d429890-5092-448b-b386-36493df22443/nova-api-api/0.log" Dec 07 10:30:42 crc kubenswrapper[4838]: I1207 10:30:42.082940 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-jtdtd_e47e7bbd-a623-4a13-ba01-2fb7ff984b40/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:43 crc kubenswrapper[4838]: I1207 10:30:43.243915 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_021e33aa-1082-46ef-afa0-1dd04aad6502/nova-metadata-log/0.log" Dec 07 10:30:43 crc kubenswrapper[4838]: I1207 10:30:43.369505 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_0f3b4db3-4d04-4198-972d-ab3722b30abe/nova-scheduler-scheduler/0.log" Dec 07 10:30:43 crc kubenswrapper[4838]: I1207 10:30:43.537397 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c/mysql-bootstrap/0.log" Dec 07 10:30:44 crc kubenswrapper[4838]: I1207 10:30:44.391909 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c/galera/0.log" Dec 07 10:30:44 crc kubenswrapper[4838]: I1207 10:30:44.465811 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_898ea16b-8db8-4d2e-9c09-d2ac3f5eaa0c/mysql-bootstrap/0.log" Dec 07 10:30:44 crc kubenswrapper[4838]: I1207 10:30:44.614667 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:30:44 crc kubenswrapper[4838]: E1207 10:30:44.614909 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:30:44 crc kubenswrapper[4838]: I1207 10:30:44.631009 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d111910-1e1a-4b37-87c2-5c8f55110a73/mysql-bootstrap/0.log" Dec 07 10:30:44 crc kubenswrapper[4838]: I1207 10:30:44.880261 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_021e33aa-1082-46ef-afa0-1dd04aad6502/nova-metadata-metadata/0.log" Dec 07 10:30:44 crc kubenswrapper[4838]: I1207 10:30:44.921387 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d111910-1e1a-4b37-87c2-5c8f55110a73/mysql-bootstrap/0.log" Dec 07 10:30:44 crc kubenswrapper[4838]: I1207 10:30:44.964760 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d111910-1e1a-4b37-87c2-5c8f55110a73/galera/0.log" Dec 07 10:30:45 crc kubenswrapper[4838]: I1207 10:30:45.124724 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_096a9643-43b6-4f09-8be1-cd739058c12e/openstackclient/0.log" Dec 07 10:30:45 crc kubenswrapper[4838]: I1207 10:30:45.287928 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-j7t4b_e7b95347-0585-4016-a3fa-a17c6d9042a3/openstack-network-exporter/0.log" Dec 07 10:30:45 crc kubenswrapper[4838]: I1207 10:30:45.355980 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovsdb-server-init/0.log" Dec 07 10:30:45 crc kubenswrapper[4838]: I1207 10:30:45.626488 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovsdb-server-init/0.log" Dec 07 10:30:45 crc kubenswrapper[4838]: I1207 10:30:45.656528 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovs-vswitchd/0.log" Dec 07 10:30:45 crc kubenswrapper[4838]: I1207 10:30:45.744766 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gb65g_dc745e13-8d6d-4e12-a733-5c081c0e4283/ovsdb-server/0.log" Dec 07 10:30:45 crc kubenswrapper[4838]: I1207 10:30:45.913980 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-sbl7t_a3ae58f2-6cd2-4782-96b1-a88597b25b37/ovn-controller/0.log" Dec 07 10:30:46 crc kubenswrapper[4838]: I1207 10:30:46.085249 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-84r6c_6ae7a9e2-889e-499b-8f64-d38976e8d429/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:46 crc kubenswrapper[4838]: I1207 10:30:46.142150 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_065bed0f-e9fc-4d9a-a455-46d9d77e7bb2/openstack-network-exporter/0.log" Dec 07 10:30:46 crc kubenswrapper[4838]: I1207 10:30:46.333533 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_065bed0f-e9fc-4d9a-a455-46d9d77e7bb2/ovn-northd/0.log" Dec 07 10:30:46 crc kubenswrapper[4838]: I1207 10:30:46.459522 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cec74e7b-a2aa-46a0-b318-61decd213728/openstack-network-exporter/0.log" Dec 07 10:30:46 crc kubenswrapper[4838]: I1207 10:30:46.466839 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cec74e7b-a2aa-46a0-b318-61decd213728/ovsdbserver-nb/0.log" Dec 07 10:30:46 crc kubenswrapper[4838]: I1207 10:30:46.750725 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e06fb79a-09b7-42af-bb04-1ef9ba3f7560/openstack-network-exporter/0.log" Dec 07 10:30:46 crc kubenswrapper[4838]: I1207 10:30:46.798371 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e06fb79a-09b7-42af-bb04-1ef9ba3f7560/ovsdbserver-sb/0.log" Dec 07 10:30:47 crc kubenswrapper[4838]: I1207 10:30:47.121877 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9b574bc66-nsc2v_87509a0e-d4f9-4d24-b490-9d5e2675aa21/placement-api/0.log" Dec 07 10:30:47 crc kubenswrapper[4838]: I1207 10:30:47.122733 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a017a788-cb2e-45ed-bceb-26ea6b69a02d/setup-container/0.log" Dec 07 10:30:47 crc kubenswrapper[4838]: I1207 10:30:47.186032 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9b574bc66-nsc2v_87509a0e-d4f9-4d24-b490-9d5e2675aa21/placement-log/0.log" Dec 07 10:30:47 crc kubenswrapper[4838]: I1207 10:30:47.846394 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a017a788-cb2e-45ed-bceb-26ea6b69a02d/setup-container/0.log" Dec 07 10:30:47 crc kubenswrapper[4838]: I1207 10:30:47.950266 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a017a788-cb2e-45ed-bceb-26ea6b69a02d/rabbitmq/0.log" Dec 07 10:30:47 crc kubenswrapper[4838]: I1207 10:30:47.956117 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_688c3086-fd09-4b9d-921e-e41c7f8a01ac/setup-container/0.log" Dec 07 10:30:48 crc kubenswrapper[4838]: I1207 10:30:48.199179 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_688c3086-fd09-4b9d-921e-e41c7f8a01ac/rabbitmq/0.log" Dec 07 10:30:48 crc kubenswrapper[4838]: I1207 10:30:48.307037 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-tk56g_df0eb7df-51e9-4da6-9e69-0b11a79ad3fa/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:48 crc kubenswrapper[4838]: I1207 10:30:48.345385 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_688c3086-fd09-4b9d-921e-e41c7f8a01ac/setup-container/0.log" Dec 07 10:30:48 crc kubenswrapper[4838]: I1207 10:30:48.609089 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-d5trb_8587abfa-6182-4613-b458-69c4f1bf2eca/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:48 crc kubenswrapper[4838]: I1207 10:30:48.862016 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-d67k5_e471497e-008f-40a2-929e-33e08ea45f63/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:48 crc kubenswrapper[4838]: I1207 10:30:48.931213 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8grpd_131b49fd-19d7-4188-a131-2d684a0eee60/ssh-known-hosts-edpm-deployment/0.log" Dec 07 10:30:49 crc kubenswrapper[4838]: I1207 10:30:49.237483 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a430a791-ad62-4246-8422-6bae0b6ce772/tempest-tests-tempest-tests-runner/0.log" Dec 07 10:30:49 crc kubenswrapper[4838]: I1207 10:30:49.243347 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_17d4b7b1-e69f-444b-bd80-753b07d66af7/test-operator-logs-container/0.log" Dec 07 10:30:49 crc kubenswrapper[4838]: I1207 10:30:49.687644 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-99vbk_f9c17708-e8da-45b1-b919-354138adf40b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 10:30:56 crc kubenswrapper[4838]: I1207 10:30:56.003753 4838 scope.go:117] "RemoveContainer" containerID="3cadfad733aeffaa303627fe927a40291abae7f178500385ddfaf21783466b54" Dec 07 10:30:59 crc kubenswrapper[4838]: I1207 10:30:59.620674 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:30:59 crc kubenswrapper[4838]: E1207 10:30:59.621503 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:31:01 crc kubenswrapper[4838]: I1207 10:31:01.630330 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_76cc3a82-a620-435a-99f0-9a91297d2ee2/memcached/0.log" Dec 07 10:31:13 crc kubenswrapper[4838]: I1207 10:31:13.617334 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:31:13 crc kubenswrapper[4838]: E1207 10:31:13.618173 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:31:23 crc kubenswrapper[4838]: I1207 10:31:23.608396 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/util/0.log" Dec 07 10:31:23 crc kubenswrapper[4838]: I1207 10:31:23.776939 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/util/0.log" Dec 07 10:31:23 crc kubenswrapper[4838]: I1207 10:31:23.806081 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/pull/0.log" Dec 07 10:31:23 crc kubenswrapper[4838]: I1207 10:31:23.877695 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/pull/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.056455 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/extract/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.104203 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/util/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.120312 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b9120eefd1323238be49f2f5c049f678a31f5be734312c37cc3d85ecabl8jzs_3a7e3122-aebe-4d6b-8611-928f56c27d66/pull/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.363744 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5cgxx_2e6976c0-74c7-4f62-88e0-6f6b635f829b/manager/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.387525 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5cgxx_2e6976c0-74c7-4f62-88e0-6f6b635f829b/kube-rbac-proxy/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.443640 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-p5xdf_e9af8ce9-a7b0-4bca-b834-2f2f494eb434/kube-rbac-proxy/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.614247 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-p5xdf_e9af8ce9-a7b0-4bca-b834-2f2f494eb434/manager/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.654686 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vb8wb_c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd/manager/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.687681 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vb8wb_c9ccd4bb-b4fc-4bfd-92ec-8dffd1fc25bd/kube-rbac-proxy/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.927722 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-9hg7c_f91ec7b3-50b5-4f9f-95a0-c65abeeb879f/kube-rbac-proxy/0.log" Dec 07 10:31:24 crc kubenswrapper[4838]: I1207 10:31:24.938835 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-9hg7c_f91ec7b3-50b5-4f9f-95a0-c65abeeb879f/manager/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.098158 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4hn7r_5d086115-289a-4478-a657-7ea3927165cd/kube-rbac-proxy/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.165411 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4hn7r_5d086115-289a-4478-a657-7ea3927165cd/manager/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.208631 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2ftqb_49a96bc4-d983-4088-973e-ebb95fdcc725/kube-rbac-proxy/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.337082 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2ftqb_49a96bc4-d983-4088-973e-ebb95fdcc725/manager/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.390537 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4q7pd_fac41ae4-04ea-4e07-b672-13fdf35a6530/kube-rbac-proxy/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.683729 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4q7pd_fac41ae4-04ea-4e07-b672-13fdf35a6530/manager/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.726570 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-pmdl9_ba02a228-91ff-41cf-88ac-b027e697dff9/kube-rbac-proxy/0.log" Dec 07 10:31:25 crc kubenswrapper[4838]: I1207 10:31:25.734263 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-pmdl9_ba02a228-91ff-41cf-88ac-b027e697dff9/manager/0.log" Dec 07 10:31:26 crc kubenswrapper[4838]: I1207 10:31:26.437378 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-gqrds_23f05536-d749-43c5-a7cc-7dc63f46e288/kube-rbac-proxy/0.log" Dec 07 10:31:26 crc kubenswrapper[4838]: I1207 10:31:26.552878 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-gqrds_23f05536-d749-43c5-a7cc-7dc63f46e288/manager/0.log" Dec 07 10:31:26 crc kubenswrapper[4838]: I1207 10:31:26.611804 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-hbffh_7feb13b5-ed7a-47bf-82f1-b749d5b9574e/kube-rbac-proxy/0.log" Dec 07 10:31:26 crc kubenswrapper[4838]: I1207 10:31:26.614407 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:31:26 crc kubenswrapper[4838]: E1207 10:31:26.614709 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:31:26 crc kubenswrapper[4838]: I1207 10:31:26.796266 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6cwcf_3bd0361d-590b-4a06-b32c-6f7ceb427fbd/kube-rbac-proxy/0.log" Dec 07 10:31:26 crc kubenswrapper[4838]: I1207 10:31:26.814019 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-hbffh_7feb13b5-ed7a-47bf-82f1-b749d5b9574e/manager/0.log" Dec 07 10:31:26 crc kubenswrapper[4838]: I1207 10:31:26.901392 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6cwcf_3bd0361d-590b-4a06-b32c-6f7ceb427fbd/manager/0.log" Dec 07 10:31:27 crc kubenswrapper[4838]: I1207 10:31:27.038521 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-9cwrh_3af86462-0c95-4f05-9e30-c70da589a944/kube-rbac-proxy/0.log" Dec 07 10:31:27 crc kubenswrapper[4838]: I1207 10:31:27.109036 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-9cwrh_3af86462-0c95-4f05-9e30-c70da589a944/manager/0.log" Dec 07 10:31:27 crc kubenswrapper[4838]: I1207 10:31:27.190030 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-t6tz7_86196566-5b23-4cf1-924e-8a802086e443/kube-rbac-proxy/0.log" Dec 07 10:31:27 crc kubenswrapper[4838]: I1207 10:31:27.371232 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-t6tz7_86196566-5b23-4cf1-924e-8a802086e443/manager/0.log" Dec 07 10:31:27 crc kubenswrapper[4838]: I1207 10:31:27.474451 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-54jxx_30216693-f616-448b-b85d-1c6482317ec5/kube-rbac-proxy/0.log" Dec 07 10:31:27 crc kubenswrapper[4838]: I1207 10:31:27.545036 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-54jxx_30216693-f616-448b-b85d-1c6482317ec5/manager/0.log" Dec 07 10:31:28 crc kubenswrapper[4838]: I1207 10:31:28.147163 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-744f8cb766qxndm_07975bb2-d979-4265-bba2-1254d58b7267/kube-rbac-proxy/0.log" Dec 07 10:31:28 crc kubenswrapper[4838]: I1207 10:31:28.238349 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-744f8cb766qxndm_07975bb2-d979-4265-bba2-1254d58b7267/manager/0.log" Dec 07 10:31:28 crc kubenswrapper[4838]: I1207 10:31:28.723807 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2sblj_0851e8d2-5d7c-4e07-aed5-d743468017ec/registry-server/0.log" Dec 07 10:31:28 crc kubenswrapper[4838]: I1207 10:31:28.737999 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5f48db4cb9-cs44h_e28de98c-d75e-4673-8299-fd5abba22b44/operator/0.log" Dec 07 10:31:28 crc kubenswrapper[4838]: I1207 10:31:28.857993 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-b29lr_fdc08e45-9f41-4a0c-8478-10ab2b22ad4e/kube-rbac-proxy/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.052648 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9jtsq_5fc12298-865b-411d-ab04-6cc9d99d8030/kube-rbac-proxy/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.142174 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9jtsq_5fc12298-865b-411d-ab04-6cc9d99d8030/manager/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.278601 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-b29lr_fdc08e45-9f41-4a0c-8478-10ab2b22ad4e/manager/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.487022 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-tl2fn_700b9dd5-3c10-4fa7-bde4-0c9a4f778608/operator/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.504192 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5b47684954-7mp9q_1f077e27-60ec-4c93-aada-4a002c07a70c/manager/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.606540 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-6vrx6_5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f/kube-rbac-proxy/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.744145 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-6vrx6_5b3f1aed-fa07-4b44-a8a7-4cd24579fc2f/manager/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.757670 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-gbbg9_71940e0c-91e5-4634-92d3-d6061fbfc768/kube-rbac-proxy/0.log" Dec 07 10:31:29 crc kubenswrapper[4838]: I1207 10:31:29.866761 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-gbbg9_71940e0c-91e5-4634-92d3-d6061fbfc768/manager/0.log" Dec 07 10:31:30 crc kubenswrapper[4838]: I1207 10:31:30.035863 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wz9hz_44803609-79be-4c16-bd10-3d6820d8f24b/manager/0.log" Dec 07 10:31:30 crc kubenswrapper[4838]: I1207 10:31:30.068670 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wz9hz_44803609-79be-4c16-bd10-3d6820d8f24b/kube-rbac-proxy/0.log" Dec 07 10:31:30 crc kubenswrapper[4838]: I1207 10:31:30.197610 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-v5hzl_3c038260-fe42-4c67-8e8c-b994a83c6156/kube-rbac-proxy/0.log" Dec 07 10:31:30 crc kubenswrapper[4838]: I1207 10:31:30.207963 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-v5hzl_3c038260-fe42-4c67-8e8c-b994a83c6156/manager/0.log" Dec 07 10:31:40 crc kubenswrapper[4838]: I1207 10:31:40.615020 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:31:40 crc kubenswrapper[4838]: E1207 10:31:40.615858 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:31:55 crc kubenswrapper[4838]: I1207 10:31:55.614154 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:31:55 crc kubenswrapper[4838]: E1207 10:31:55.614929 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:31:55 crc kubenswrapper[4838]: I1207 10:31:55.660792 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pcrsf_22165712-08e9-47cc-aaab-e5da9f92a0fb/control-plane-machine-set-operator/0.log" Dec 07 10:31:56 crc kubenswrapper[4838]: I1207 10:31:56.148169 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qll5w_e432ea72-cff7-4826-8001-125443ca158a/kube-rbac-proxy/0.log" Dec 07 10:31:56 crc kubenswrapper[4838]: I1207 10:31:56.303040 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qll5w_e432ea72-cff7-4826-8001-125443ca158a/machine-api-operator/0.log" Dec 07 10:32:07 crc kubenswrapper[4838]: I1207 10:32:07.615052 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:32:07 crc kubenswrapper[4838]: E1207 10:32:07.615902 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:32:09 crc kubenswrapper[4838]: I1207 10:32:09.708205 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2h6mq_d7e66c54-9769-4287-8b1c-f1d1a56b08a1/cert-manager-controller/0.log" Dec 07 10:32:09 crc kubenswrapper[4838]: I1207 10:32:09.928426 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-6knh9_c0fb9b54-1b0a-4cab-8bad-669c11ea8a73/cert-manager-cainjector/0.log" Dec 07 10:32:09 crc kubenswrapper[4838]: I1207 10:32:09.938324 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-87mkv_72c34e1b-8df3-4622-9cdb-b44adf0e45bf/cert-manager-webhook/0.log" Dec 07 10:32:22 crc kubenswrapper[4838]: I1207 10:32:22.615137 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:32:22 crc kubenswrapper[4838]: E1207 10:32:22.615784 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:32:23 crc kubenswrapper[4838]: I1207 10:32:23.991107 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-lzwd2_013a32de-8cd7-44da-b292-5948ab0bb6c7/nmstate-console-plugin/0.log" Dec 07 10:32:24 crc kubenswrapper[4838]: I1207 10:32:24.716448 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-twb76_a130a77c-66e9-48b2-a9c9-36c44b7b8eec/nmstate-handler/0.log" Dec 07 10:32:24 crc kubenswrapper[4838]: I1207 10:32:24.738959 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p58hh_11834459-a487-4c2f-9521-add0e1726e9d/kube-rbac-proxy/0.log" Dec 07 10:32:24 crc kubenswrapper[4838]: I1207 10:32:24.778503 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p58hh_11834459-a487-4c2f-9521-add0e1726e9d/nmstate-metrics/0.log" Dec 07 10:32:24 crc kubenswrapper[4838]: I1207 10:32:24.975408 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-4hcdm_e1df0bd8-ff98-4094-92b6-23f266883f8a/nmstate-operator/0.log" Dec 07 10:32:25 crc kubenswrapper[4838]: I1207 10:32:25.033781 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-jcgsg_9d7f7de9-0d6e-4bf5-83c0-5f703cb433af/nmstate-webhook/0.log" Dec 07 10:32:34 crc kubenswrapper[4838]: I1207 10:32:34.615684 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:32:34 crc kubenswrapper[4838]: E1207 10:32:34.616887 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:32:41 crc kubenswrapper[4838]: I1207 10:32:41.820288 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj79s_4fa3fe39-d052-4d6f-881c-526b036b4bbf/kube-rbac-proxy/0.log" Dec 07 10:32:41 crc kubenswrapper[4838]: I1207 10:32:41.902762 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj79s_4fa3fe39-d052-4d6f-881c-526b036b4bbf/controller/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.113590 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-6btm8_39164f82-bb66-4fe1-a4e4-0007d9697e40/frr-k8s-webhook-server/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.197025 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.388032 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.389416 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.427064 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.435168 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.643344 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.651519 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.687172 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.708652 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.887775 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-frr-files/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.890000 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-metrics/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.892510 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/cp-reloader/0.log" Dec 07 10:32:42 crc kubenswrapper[4838]: I1207 10:32:42.945452 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/controller/0.log" Dec 07 10:32:43 crc kubenswrapper[4838]: I1207 10:32:43.118238 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/frr-metrics/0.log" Dec 07 10:32:43 crc kubenswrapper[4838]: I1207 10:32:43.129606 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/kube-rbac-proxy/0.log" Dec 07 10:32:43 crc kubenswrapper[4838]: I1207 10:32:43.167115 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/kube-rbac-proxy-frr/0.log" Dec 07 10:32:43 crc kubenswrapper[4838]: I1207 10:32:43.829321 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/reloader/0.log" Dec 07 10:32:43 crc kubenswrapper[4838]: I1207 10:32:43.932045 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d7fb758f7-hm2jk_53c25e7b-09e8-435d-933b-9cbbe46f5ac0/manager/0.log" Dec 07 10:32:44 crc kubenswrapper[4838]: I1207 10:32:44.171002 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-94cfbc768-26mzl_cda6c003-8725-4f92-a8a8-6dad25837d56/webhook-server/0.log" Dec 07 10:32:44 crc kubenswrapper[4838]: I1207 10:32:44.255242 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x5pk6_1ef88d4e-7dae-4457-85a4-0c065afd26ad/frr/0.log" Dec 07 10:32:44 crc kubenswrapper[4838]: I1207 10:32:44.353101 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vdgqx_a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86/kube-rbac-proxy/0.log" Dec 07 10:32:44 crc kubenswrapper[4838]: I1207 10:32:44.675188 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vdgqx_a6564a93-33f2-4ae7-9ebd-fabf8dc0ce86/speaker/0.log" Dec 07 10:32:49 crc kubenswrapper[4838]: I1207 10:32:49.615578 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:32:49 crc kubenswrapper[4838]: E1207 10:32:49.616245 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:32:58 crc kubenswrapper[4838]: I1207 10:32:58.652793 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/util/0.log" Dec 07 10:32:58 crc kubenswrapper[4838]: I1207 10:32:58.878181 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/pull/0.log" Dec 07 10:32:58 crc kubenswrapper[4838]: I1207 10:32:58.910414 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/pull/0.log" Dec 07 10:32:58 crc kubenswrapper[4838]: I1207 10:32:58.939520 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/util/0.log" Dec 07 10:32:59 crc kubenswrapper[4838]: I1207 10:32:59.152410 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/util/0.log" Dec 07 10:32:59 crc kubenswrapper[4838]: I1207 10:32:59.172965 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/pull/0.log" Dec 07 10:32:59 crc kubenswrapper[4838]: I1207 10:32:59.183715 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fv5dwl_1487e6d4-4c48-4b9a-a814-fa6c0431ae7c/extract/0.log" Dec 07 10:32:59 crc kubenswrapper[4838]: I1207 10:32:59.353431 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/util/0.log" Dec 07 10:32:59 crc kubenswrapper[4838]: I1207 10:32:59.812036 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/pull/0.log" Dec 07 10:32:59 crc kubenswrapper[4838]: I1207 10:32:59.871583 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/util/0.log" Dec 07 10:32:59 crc kubenswrapper[4838]: I1207 10:32:59.879469 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/pull/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.083421 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/util/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.135404 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/extract/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.153873 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rjdcd_36061897-0432-4e69-add8-2a3560a583b5/pull/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.319211 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-utilities/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.496079 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-content/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.500392 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-utilities/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.533255 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-content/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.679848 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-utilities/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.703766 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/extract-content/0.log" Dec 07 10:33:00 crc kubenswrapper[4838]: I1207 10:33:00.895933 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-utilities/0.log" Dec 07 10:33:01 crc kubenswrapper[4838]: I1207 10:33:01.272035 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qjfhk_5989bb7b-629b-47c9-82ae-4e263e313eab/registry-server/0.log" Dec 07 10:33:01 crc kubenswrapper[4838]: I1207 10:33:01.615027 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:33:01 crc kubenswrapper[4838]: E1207 10:33:01.615273 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:33:01 crc kubenswrapper[4838]: I1207 10:33:01.713575 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-utilities/0.log" Dec 07 10:33:01 crc kubenswrapper[4838]: I1207 10:33:01.720775 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-content/0.log" Dec 07 10:33:01 crc kubenswrapper[4838]: I1207 10:33:01.753569 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-content/0.log" Dec 07 10:33:01 crc kubenswrapper[4838]: I1207 10:33:01.921661 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-utilities/0.log" Dec 07 10:33:01 crc kubenswrapper[4838]: I1207 10:33:01.941540 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/extract-content/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.222654 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-69xn8_c21136ea-de51-48f3-b79f-493f2d88ece5/marketplace-operator/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.294803 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-utilities/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.520442 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7gwg6_f62b2408-ca4c-491d-bd9d-335cdd5a5843/registry-server/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.577384 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-utilities/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.627668 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-content/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.632568 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-content/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.778801 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-utilities/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.778969 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/extract-content/0.log" Dec 07 10:33:02 crc kubenswrapper[4838]: I1207 10:33:02.855072 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-utilities/0.log" Dec 07 10:33:03 crc kubenswrapper[4838]: I1207 10:33:03.040578 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dwwqp_6ba85040-ab6e-46d8-b6ce-ea49e3a3311c/registry-server/0.log" Dec 07 10:33:03 crc kubenswrapper[4838]: I1207 10:33:03.092977 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-utilities/0.log" Dec 07 10:33:03 crc kubenswrapper[4838]: I1207 10:33:03.098767 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-content/0.log" Dec 07 10:33:03 crc kubenswrapper[4838]: I1207 10:33:03.124193 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-content/0.log" Dec 07 10:33:03 crc kubenswrapper[4838]: I1207 10:33:03.259432 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-utilities/0.log" Dec 07 10:33:03 crc kubenswrapper[4838]: I1207 10:33:03.303273 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/extract-content/0.log" Dec 07 10:33:03 crc kubenswrapper[4838]: I1207 10:33:03.918266 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pgsb_e7fa6b9c-1255-4cd3-98cb-49600d7f7668/registry-server/0.log" Dec 07 10:33:14 crc kubenswrapper[4838]: I1207 10:33:14.614193 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:33:14 crc kubenswrapper[4838]: E1207 10:33:14.615055 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:33:29 crc kubenswrapper[4838]: I1207 10:33:29.619521 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:33:29 crc kubenswrapper[4838]: E1207 10:33:29.620428 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:33:44 crc kubenswrapper[4838]: I1207 10:33:44.615458 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:33:44 crc kubenswrapper[4838]: E1207 10:33:44.616139 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:33:56 crc kubenswrapper[4838]: I1207 10:33:56.614781 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:33:56 crc kubenswrapper[4838]: E1207 10:33:56.615776 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:34:09 crc kubenswrapper[4838]: I1207 10:34:09.617279 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:34:09 crc kubenswrapper[4838]: E1207 10:34:09.621855 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:34:22 crc kubenswrapper[4838]: I1207 10:34:22.614944 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:34:22 crc kubenswrapper[4838]: E1207 10:34:22.615845 4838 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7v9qs_openshift-machine-config-operator(d1304517-3ac2-4aa3-afe2-01e27f87d880)\"" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" Dec 07 10:34:26 crc kubenswrapper[4838]: I1207 10:34:26.917182 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wvtqd"] Dec 07 10:34:26 crc kubenswrapper[4838]: E1207 10:34:26.918115 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96023c1b-8bce-4ab1-b5f2-458ea9113367" containerName="collect-profiles" Dec 07 10:34:26 crc kubenswrapper[4838]: I1207 10:34:26.918133 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="96023c1b-8bce-4ab1-b5f2-458ea9113367" containerName="collect-profiles" Dec 07 10:34:26 crc kubenswrapper[4838]: I1207 10:34:26.918389 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="96023c1b-8bce-4ab1-b5f2-458ea9113367" containerName="collect-profiles" Dec 07 10:34:26 crc kubenswrapper[4838]: I1207 10:34:26.923281 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:26 crc kubenswrapper[4838]: I1207 10:34:26.928738 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvtqd"] Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.024593 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srqw8\" (UniqueName: \"kubernetes.io/projected/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-kube-api-access-srqw8\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.025190 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-catalog-content\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.025368 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-utilities\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.127614 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-catalog-content\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.127689 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-utilities\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.127775 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srqw8\" (UniqueName: \"kubernetes.io/projected/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-kube-api-access-srqw8\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.128331 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-catalog-content\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.128421 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-utilities\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.152048 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srqw8\" (UniqueName: \"kubernetes.io/projected/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-kube-api-access-srqw8\") pod \"redhat-marketplace-wvtqd\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.242941 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:27 crc kubenswrapper[4838]: I1207 10:34:27.818097 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvtqd"] Dec 07 10:34:28 crc kubenswrapper[4838]: I1207 10:34:28.324364 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvtqd" event={"ID":"f084cdef-867d-4ca1-ae0a-fc4b23ddd059","Type":"ContainerStarted","Data":"48bf79d78de292db35bd46e1ad981239d0a2293d10b83d29f3f9637b488fcae8"} Dec 07 10:34:29 crc kubenswrapper[4838]: I1207 10:34:29.340877 4838 generic.go:334] "Generic (PLEG): container finished" podID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerID="dc4db3769161188843ad5b49749549a36cd175a9ab813bb0d6aceb92ff566181" exitCode=0 Dec 07 10:34:29 crc kubenswrapper[4838]: I1207 10:34:29.340933 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvtqd" event={"ID":"f084cdef-867d-4ca1-ae0a-fc4b23ddd059","Type":"ContainerDied","Data":"dc4db3769161188843ad5b49749549a36cd175a9ab813bb0d6aceb92ff566181"} Dec 07 10:34:29 crc kubenswrapper[4838]: I1207 10:34:29.348486 4838 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 10:34:31 crc kubenswrapper[4838]: I1207 10:34:31.359658 4838 generic.go:334] "Generic (PLEG): container finished" podID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerID="1056d601a420801fc24e4beb5e7dd1c6edd2f0561c9bbe73235138e03678d915" exitCode=0 Dec 07 10:34:31 crc kubenswrapper[4838]: I1207 10:34:31.359871 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvtqd" event={"ID":"f084cdef-867d-4ca1-ae0a-fc4b23ddd059","Type":"ContainerDied","Data":"1056d601a420801fc24e4beb5e7dd1c6edd2f0561c9bbe73235138e03678d915"} Dec 07 10:34:32 crc kubenswrapper[4838]: I1207 10:34:32.369748 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvtqd" event={"ID":"f084cdef-867d-4ca1-ae0a-fc4b23ddd059","Type":"ContainerStarted","Data":"195e14e824048b0cdf2e4e6da598286c4a43e0ca8eb0eef701c1cabd02e52242"} Dec 07 10:34:32 crc kubenswrapper[4838]: I1207 10:34:32.390601 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wvtqd" podStartSLOduration=3.9393036390000002 podStartE2EDuration="6.390584938s" podCreationTimestamp="2025-12-07 10:34:26 +0000 UTC" firstStartedPulling="2025-12-07 10:34:29.348134183 +0000 UTC m=+5286.055453220" lastFinishedPulling="2025-12-07 10:34:31.799415482 +0000 UTC m=+5288.506734519" observedRunningTime="2025-12-07 10:34:32.383074527 +0000 UTC m=+5289.090393544" watchObservedRunningTime="2025-12-07 10:34:32.390584938 +0000 UTC m=+5289.097903955" Dec 07 10:34:35 crc kubenswrapper[4838]: I1207 10:34:35.617178 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:34:36 crc kubenswrapper[4838]: I1207 10:34:36.408156 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"a2b7f561d1cd9d220ecd241c8416d4d853908ea0e5717620a2f7d502679e5af8"} Dec 07 10:34:37 crc kubenswrapper[4838]: I1207 10:34:37.243624 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:37 crc kubenswrapper[4838]: I1207 10:34:37.244774 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:37 crc kubenswrapper[4838]: I1207 10:34:37.311113 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:37 crc kubenswrapper[4838]: I1207 10:34:37.492686 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:37 crc kubenswrapper[4838]: I1207 10:34:37.556365 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvtqd"] Dec 07 10:34:39 crc kubenswrapper[4838]: I1207 10:34:39.436631 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wvtqd" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="registry-server" containerID="cri-o://195e14e824048b0cdf2e4e6da598286c4a43e0ca8eb0eef701c1cabd02e52242" gracePeriod=2 Dec 07 10:34:40 crc kubenswrapper[4838]: I1207 10:34:40.448789 4838 generic.go:334] "Generic (PLEG): container finished" podID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerID="195e14e824048b0cdf2e4e6da598286c4a43e0ca8eb0eef701c1cabd02e52242" exitCode=0 Dec 07 10:34:40 crc kubenswrapper[4838]: I1207 10:34:40.448973 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvtqd" event={"ID":"f084cdef-867d-4ca1-ae0a-fc4b23ddd059","Type":"ContainerDied","Data":"195e14e824048b0cdf2e4e6da598286c4a43e0ca8eb0eef701c1cabd02e52242"} Dec 07 10:34:40 crc kubenswrapper[4838]: I1207 10:34:40.841319 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.042056 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-utilities\") pod \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.042221 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-catalog-content\") pod \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.042540 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srqw8\" (UniqueName: \"kubernetes.io/projected/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-kube-api-access-srqw8\") pod \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\" (UID: \"f084cdef-867d-4ca1-ae0a-fc4b23ddd059\") " Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.042933 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-utilities" (OuterVolumeSpecName: "utilities") pod "f084cdef-867d-4ca1-ae0a-fc4b23ddd059" (UID: "f084cdef-867d-4ca1-ae0a-fc4b23ddd059"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.043727 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.062315 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-kube-api-access-srqw8" (OuterVolumeSpecName: "kube-api-access-srqw8") pod "f084cdef-867d-4ca1-ae0a-fc4b23ddd059" (UID: "f084cdef-867d-4ca1-ae0a-fc4b23ddd059"). InnerVolumeSpecName "kube-api-access-srqw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.069237 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f084cdef-867d-4ca1-ae0a-fc4b23ddd059" (UID: "f084cdef-867d-4ca1-ae0a-fc4b23ddd059"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.145295 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srqw8\" (UniqueName: \"kubernetes.io/projected/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-kube-api-access-srqw8\") on node \"crc\" DevicePath \"\"" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.145636 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f084cdef-867d-4ca1-ae0a-fc4b23ddd059-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.469499 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvtqd" event={"ID":"f084cdef-867d-4ca1-ae0a-fc4b23ddd059","Type":"ContainerDied","Data":"48bf79d78de292db35bd46e1ad981239d0a2293d10b83d29f3f9637b488fcae8"} Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.469560 4838 scope.go:117] "RemoveContainer" containerID="195e14e824048b0cdf2e4e6da598286c4a43e0ca8eb0eef701c1cabd02e52242" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.469588 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvtqd" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.507912 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvtqd"] Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.516042 4838 scope.go:117] "RemoveContainer" containerID="1056d601a420801fc24e4beb5e7dd1c6edd2f0561c9bbe73235138e03678d915" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.517368 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvtqd"] Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.633259 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" path="/var/lib/kubelet/pods/f084cdef-867d-4ca1-ae0a-fc4b23ddd059/volumes" Dec 07 10:34:41 crc kubenswrapper[4838]: I1207 10:34:41.844418 4838 scope.go:117] "RemoveContainer" containerID="dc4db3769161188843ad5b49749549a36cd175a9ab813bb0d6aceb92ff566181" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.315787 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cjd27"] Dec 07 10:34:47 crc kubenswrapper[4838]: E1207 10:34:47.317070 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="extract-content" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.317091 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="extract-content" Dec 07 10:34:47 crc kubenswrapper[4838]: E1207 10:34:47.317143 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="extract-utilities" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.317152 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="extract-utilities" Dec 07 10:34:47 crc kubenswrapper[4838]: E1207 10:34:47.317178 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="registry-server" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.317188 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="registry-server" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.317404 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="f084cdef-867d-4ca1-ae0a-fc4b23ddd059" containerName="registry-server" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.321776 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.340474 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cjd27"] Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.389540 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-utilities\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.389680 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdfkc\" (UniqueName: \"kubernetes.io/projected/cbb343dc-9fc6-4e07-8e59-624153b1353d-kube-api-access-hdfkc\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.389723 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-catalog-content\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.490969 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-utilities\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.491106 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdfkc\" (UniqueName: \"kubernetes.io/projected/cbb343dc-9fc6-4e07-8e59-624153b1353d-kube-api-access-hdfkc\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.491146 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-catalog-content\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.491513 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-utilities\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.491554 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-catalog-content\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.513937 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdfkc\" (UniqueName: \"kubernetes.io/projected/cbb343dc-9fc6-4e07-8e59-624153b1353d-kube-api-access-hdfkc\") pod \"redhat-operators-cjd27\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:47 crc kubenswrapper[4838]: I1207 10:34:47.642044 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:48 crc kubenswrapper[4838]: I1207 10:34:48.115935 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cjd27"] Dec 07 10:34:48 crc kubenswrapper[4838]: I1207 10:34:48.549219 4838 generic.go:334] "Generic (PLEG): container finished" podID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerID="0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92" exitCode=0 Dec 07 10:34:48 crc kubenswrapper[4838]: I1207 10:34:48.549496 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cjd27" event={"ID":"cbb343dc-9fc6-4e07-8e59-624153b1353d","Type":"ContainerDied","Data":"0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92"} Dec 07 10:34:48 crc kubenswrapper[4838]: I1207 10:34:48.549522 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cjd27" event={"ID":"cbb343dc-9fc6-4e07-8e59-624153b1353d","Type":"ContainerStarted","Data":"72d6fb221b23282287f1a81c3b458ab52cfe04d16902ef526e68d975f572bb32"} Dec 07 10:34:49 crc kubenswrapper[4838]: I1207 10:34:49.566865 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cjd27" event={"ID":"cbb343dc-9fc6-4e07-8e59-624153b1353d","Type":"ContainerStarted","Data":"1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18"} Dec 07 10:34:52 crc kubenswrapper[4838]: I1207 10:34:52.618757 4838 generic.go:334] "Generic (PLEG): container finished" podID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerID="1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18" exitCode=0 Dec 07 10:34:52 crc kubenswrapper[4838]: I1207 10:34:52.618851 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cjd27" event={"ID":"cbb343dc-9fc6-4e07-8e59-624153b1353d","Type":"ContainerDied","Data":"1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18"} Dec 07 10:34:53 crc kubenswrapper[4838]: I1207 10:34:53.633478 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cjd27" event={"ID":"cbb343dc-9fc6-4e07-8e59-624153b1353d","Type":"ContainerStarted","Data":"2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052"} Dec 07 10:34:53 crc kubenswrapper[4838]: I1207 10:34:53.661359 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cjd27" podStartSLOduration=2.169277314 podStartE2EDuration="6.661339712s" podCreationTimestamp="2025-12-07 10:34:47 +0000 UTC" firstStartedPulling="2025-12-07 10:34:48.561024169 +0000 UTC m=+5305.268343186" lastFinishedPulling="2025-12-07 10:34:53.053086567 +0000 UTC m=+5309.760405584" observedRunningTime="2025-12-07 10:34:53.658579755 +0000 UTC m=+5310.365898772" watchObservedRunningTime="2025-12-07 10:34:53.661339712 +0000 UTC m=+5310.368658729" Dec 07 10:34:57 crc kubenswrapper[4838]: I1207 10:34:57.643005 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:57 crc kubenswrapper[4838]: I1207 10:34:57.643673 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:34:58 crc kubenswrapper[4838]: I1207 10:34:58.720330 4838 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cjd27" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="registry-server" probeResult="failure" output=< Dec 07 10:34:58 crc kubenswrapper[4838]: timeout: failed to connect service ":50051" within 1s Dec 07 10:34:58 crc kubenswrapper[4838]: > Dec 07 10:35:07 crc kubenswrapper[4838]: I1207 10:35:07.715052 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:35:07 crc kubenswrapper[4838]: I1207 10:35:07.781064 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:35:07 crc kubenswrapper[4838]: I1207 10:35:07.961082 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cjd27"] Dec 07 10:35:08 crc kubenswrapper[4838]: I1207 10:35:08.811569 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cjd27" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="registry-server" containerID="cri-o://2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052" gracePeriod=2 Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.308192 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.438413 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-catalog-content\") pod \"cbb343dc-9fc6-4e07-8e59-624153b1353d\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.438873 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdfkc\" (UniqueName: \"kubernetes.io/projected/cbb343dc-9fc6-4e07-8e59-624153b1353d-kube-api-access-hdfkc\") pod \"cbb343dc-9fc6-4e07-8e59-624153b1353d\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.439064 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-utilities\") pod \"cbb343dc-9fc6-4e07-8e59-624153b1353d\" (UID: \"cbb343dc-9fc6-4e07-8e59-624153b1353d\") " Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.441212 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-utilities" (OuterVolumeSpecName: "utilities") pod "cbb343dc-9fc6-4e07-8e59-624153b1353d" (UID: "cbb343dc-9fc6-4e07-8e59-624153b1353d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.445786 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb343dc-9fc6-4e07-8e59-624153b1353d-kube-api-access-hdfkc" (OuterVolumeSpecName: "kube-api-access-hdfkc") pod "cbb343dc-9fc6-4e07-8e59-624153b1353d" (UID: "cbb343dc-9fc6-4e07-8e59-624153b1353d"). InnerVolumeSpecName "kube-api-access-hdfkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.525270 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbb343dc-9fc6-4e07-8e59-624153b1353d" (UID: "cbb343dc-9fc6-4e07-8e59-624153b1353d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.541043 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdfkc\" (UniqueName: \"kubernetes.io/projected/cbb343dc-9fc6-4e07-8e59-624153b1353d-kube-api-access-hdfkc\") on node \"crc\" DevicePath \"\"" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.541081 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.541117 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb343dc-9fc6-4e07-8e59-624153b1353d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.825623 4838 generic.go:334] "Generic (PLEG): container finished" podID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerID="2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052" exitCode=0 Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.825670 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cjd27" event={"ID":"cbb343dc-9fc6-4e07-8e59-624153b1353d","Type":"ContainerDied","Data":"2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052"} Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.825708 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cjd27" event={"ID":"cbb343dc-9fc6-4e07-8e59-624153b1353d","Type":"ContainerDied","Data":"72d6fb221b23282287f1a81c3b458ab52cfe04d16902ef526e68d975f572bb32"} Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.825712 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cjd27" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.825728 4838 scope.go:117] "RemoveContainer" containerID="2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.861613 4838 scope.go:117] "RemoveContainer" containerID="1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.868160 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cjd27"] Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.877721 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cjd27"] Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.895573 4838 scope.go:117] "RemoveContainer" containerID="0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.947565 4838 scope.go:117] "RemoveContainer" containerID="2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052" Dec 07 10:35:09 crc kubenswrapper[4838]: E1207 10:35:09.948057 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052\": container with ID starting with 2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052 not found: ID does not exist" containerID="2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.948103 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052"} err="failed to get container status \"2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052\": rpc error: code = NotFound desc = could not find container \"2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052\": container with ID starting with 2ec21c07c775627412a422ccf353d8ca456e92c37be329cc8c157e4b9a9a8052 not found: ID does not exist" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.948128 4838 scope.go:117] "RemoveContainer" containerID="1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18" Dec 07 10:35:09 crc kubenswrapper[4838]: E1207 10:35:09.948442 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18\": container with ID starting with 1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18 not found: ID does not exist" containerID="1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.948470 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18"} err="failed to get container status \"1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18\": rpc error: code = NotFound desc = could not find container \"1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18\": container with ID starting with 1bb5668a9107b3e405a96421273ab7ff49dca567750b01cc1c838a9c7ba46d18 not found: ID does not exist" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.948488 4838 scope.go:117] "RemoveContainer" containerID="0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92" Dec 07 10:35:09 crc kubenswrapper[4838]: E1207 10:35:09.948712 4838 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92\": container with ID starting with 0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92 not found: ID does not exist" containerID="0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92" Dec 07 10:35:09 crc kubenswrapper[4838]: I1207 10:35:09.948744 4838 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92"} err="failed to get container status \"0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92\": rpc error: code = NotFound desc = could not find container \"0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92\": container with ID starting with 0918e6e4c374dfc1aa2269ad8bb95e46a33cbac9233f7c0a8b4054cf986e7d92 not found: ID does not exist" Dec 07 10:35:11 crc kubenswrapper[4838]: I1207 10:35:11.633378 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" path="/var/lib/kubelet/pods/cbb343dc-9fc6-4e07-8e59-624153b1353d/volumes" Dec 07 10:35:24 crc kubenswrapper[4838]: I1207 10:35:24.976155 4838 generic.go:334] "Generic (PLEG): container finished" podID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerID="73208fc40e1d34e46ee779e0e6d0364ace68556c65f4cfb1d2db33ce85a7c12e" exitCode=0 Dec 07 10:35:24 crc kubenswrapper[4838]: I1207 10:35:24.976249 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttxz8/must-gather-wttxv" event={"ID":"3766e4fb-8a36-431a-9cfe-8984b5d60f3b","Type":"ContainerDied","Data":"73208fc40e1d34e46ee779e0e6d0364ace68556c65f4cfb1d2db33ce85a7c12e"} Dec 07 10:35:24 crc kubenswrapper[4838]: I1207 10:35:24.978447 4838 scope.go:117] "RemoveContainer" containerID="73208fc40e1d34e46ee779e0e6d0364ace68556c65f4cfb1d2db33ce85a7c12e" Dec 07 10:35:25 crc kubenswrapper[4838]: I1207 10:35:25.680366 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ttxz8_must-gather-wttxv_3766e4fb-8a36-431a-9cfe-8984b5d60f3b/gather/0.log" Dec 07 10:35:38 crc kubenswrapper[4838]: I1207 10:35:38.830269 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ttxz8/must-gather-wttxv"] Dec 07 10:35:38 crc kubenswrapper[4838]: I1207 10:35:38.831115 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ttxz8/must-gather-wttxv" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerName="copy" containerID="cri-o://357dbe7a0d2abba6795b387e7e1a7d04cbcadf0b9c48d0627a326af3ea99460a" gracePeriod=2 Dec 07 10:35:38 crc kubenswrapper[4838]: I1207 10:35:38.842216 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ttxz8/must-gather-wttxv"] Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.138434 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ttxz8_must-gather-wttxv_3766e4fb-8a36-431a-9cfe-8984b5d60f3b/copy/0.log" Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.139656 4838 generic.go:334] "Generic (PLEG): container finished" podID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerID="357dbe7a0d2abba6795b387e7e1a7d04cbcadf0b9c48d0627a326af3ea99460a" exitCode=143 Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.265680 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ttxz8_must-gather-wttxv_3766e4fb-8a36-431a-9cfe-8984b5d60f3b/copy/0.log" Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.266971 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.323681 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-must-gather-output\") pod \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.323917 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shll6\" (UniqueName: \"kubernetes.io/projected/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-kube-api-access-shll6\") pod \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\" (UID: \"3766e4fb-8a36-431a-9cfe-8984b5d60f3b\") " Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.335040 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-kube-api-access-shll6" (OuterVolumeSpecName: "kube-api-access-shll6") pod "3766e4fb-8a36-431a-9cfe-8984b5d60f3b" (UID: "3766e4fb-8a36-431a-9cfe-8984b5d60f3b"). InnerVolumeSpecName "kube-api-access-shll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.427006 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shll6\" (UniqueName: \"kubernetes.io/projected/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-kube-api-access-shll6\") on node \"crc\" DevicePath \"\"" Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.506914 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3766e4fb-8a36-431a-9cfe-8984b5d60f3b" (UID: "3766e4fb-8a36-431a-9cfe-8984b5d60f3b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.528855 4838 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3766e4fb-8a36-431a-9cfe-8984b5d60f3b-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 07 10:35:39 crc kubenswrapper[4838]: I1207 10:35:39.624893 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" path="/var/lib/kubelet/pods/3766e4fb-8a36-431a-9cfe-8984b5d60f3b/volumes" Dec 07 10:35:40 crc kubenswrapper[4838]: I1207 10:35:40.149319 4838 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ttxz8_must-gather-wttxv_3766e4fb-8a36-431a-9cfe-8984b5d60f3b/copy/0.log" Dec 07 10:35:40 crc kubenswrapper[4838]: I1207 10:35:40.150523 4838 scope.go:117] "RemoveContainer" containerID="357dbe7a0d2abba6795b387e7e1a7d04cbcadf0b9c48d0627a326af3ea99460a" Dec 07 10:35:40 crc kubenswrapper[4838]: I1207 10:35:40.150859 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttxz8/must-gather-wttxv" Dec 07 10:35:40 crc kubenswrapper[4838]: I1207 10:35:40.170065 4838 scope.go:117] "RemoveContainer" containerID="73208fc40e1d34e46ee779e0e6d0364ace68556c65f4cfb1d2db33ce85a7c12e" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.716262 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c47cd"] Dec 07 10:35:57 crc kubenswrapper[4838]: E1207 10:35:57.719387 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="registry-server" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719408 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="registry-server" Dec 07 10:35:57 crc kubenswrapper[4838]: E1207 10:35:57.719440 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerName="copy" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719448 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerName="copy" Dec 07 10:35:57 crc kubenswrapper[4838]: E1207 10:35:57.719476 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="extract-content" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719485 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="extract-content" Dec 07 10:35:57 crc kubenswrapper[4838]: E1207 10:35:57.719502 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="extract-utilities" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719510 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="extract-utilities" Dec 07 10:35:57 crc kubenswrapper[4838]: E1207 10:35:57.719548 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerName="gather" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719555 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerName="gather" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719763 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerName="gather" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719777 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb343dc-9fc6-4e07-8e59-624153b1353d" containerName="registry-server" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.719797 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="3766e4fb-8a36-431a-9cfe-8984b5d60f3b" containerName="copy" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.721500 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.739230 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c47cd"] Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.792405 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-utilities\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.792479 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-catalog-content\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.792603 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v57h4\" (UniqueName: \"kubernetes.io/projected/773fe600-962c-4dc8-acea-b7a25a2a3c38-kube-api-access-v57h4\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.894285 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v57h4\" (UniqueName: \"kubernetes.io/projected/773fe600-962c-4dc8-acea-b7a25a2a3c38-kube-api-access-v57h4\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.894604 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-utilities\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.894706 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-catalog-content\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.895051 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-utilities\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.895179 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-catalog-content\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:57 crc kubenswrapper[4838]: I1207 10:35:57.915973 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v57h4\" (UniqueName: \"kubernetes.io/projected/773fe600-962c-4dc8-acea-b7a25a2a3c38-kube-api-access-v57h4\") pod \"certified-operators-c47cd\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:58 crc kubenswrapper[4838]: I1207 10:35:58.039092 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:35:58 crc kubenswrapper[4838]: I1207 10:35:58.595249 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c47cd"] Dec 07 10:35:59 crc kubenswrapper[4838]: I1207 10:35:59.339215 4838 generic.go:334] "Generic (PLEG): container finished" podID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerID="2bbcff80c35cb726643d5665c28f94e6da95921329807f8a1404b1ae4470d229" exitCode=0 Dec 07 10:35:59 crc kubenswrapper[4838]: I1207 10:35:59.339257 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47cd" event={"ID":"773fe600-962c-4dc8-acea-b7a25a2a3c38","Type":"ContainerDied","Data":"2bbcff80c35cb726643d5665c28f94e6da95921329807f8a1404b1ae4470d229"} Dec 07 10:35:59 crc kubenswrapper[4838]: I1207 10:35:59.339468 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47cd" event={"ID":"773fe600-962c-4dc8-acea-b7a25a2a3c38","Type":"ContainerStarted","Data":"480b5a024297248a7812b352e2e48bb556e1933d32731096630e9da5ba84104e"} Dec 07 10:36:00 crc kubenswrapper[4838]: I1207 10:36:00.352852 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47cd" event={"ID":"773fe600-962c-4dc8-acea-b7a25a2a3c38","Type":"ContainerStarted","Data":"01305b009a24c781dfe4960ebf546ed581bf4467c29250d12d8e2ffa255232be"} Dec 07 10:36:01 crc kubenswrapper[4838]: I1207 10:36:01.365380 4838 generic.go:334] "Generic (PLEG): container finished" podID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerID="01305b009a24c781dfe4960ebf546ed581bf4467c29250d12d8e2ffa255232be" exitCode=0 Dec 07 10:36:01 crc kubenswrapper[4838]: I1207 10:36:01.365448 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47cd" event={"ID":"773fe600-962c-4dc8-acea-b7a25a2a3c38","Type":"ContainerDied","Data":"01305b009a24c781dfe4960ebf546ed581bf4467c29250d12d8e2ffa255232be"} Dec 07 10:36:02 crc kubenswrapper[4838]: I1207 10:36:02.376063 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47cd" event={"ID":"773fe600-962c-4dc8-acea-b7a25a2a3c38","Type":"ContainerStarted","Data":"cc11e4783aba1b2c63333fd803a9796f86fbcea2a2de4b4c634c49665280bae1"} Dec 07 10:36:02 crc kubenswrapper[4838]: I1207 10:36:02.405156 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c47cd" podStartSLOduration=2.974685701 podStartE2EDuration="5.40513781s" podCreationTimestamp="2025-12-07 10:35:57 +0000 UTC" firstStartedPulling="2025-12-07 10:35:59.343201408 +0000 UTC m=+5376.050520425" lastFinishedPulling="2025-12-07 10:36:01.773653477 +0000 UTC m=+5378.480972534" observedRunningTime="2025-12-07 10:36:02.400026185 +0000 UTC m=+5379.107345212" watchObservedRunningTime="2025-12-07 10:36:02.40513781 +0000 UTC m=+5379.112456837" Dec 07 10:36:08 crc kubenswrapper[4838]: I1207 10:36:08.040401 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:36:08 crc kubenswrapper[4838]: I1207 10:36:08.041175 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:36:08 crc kubenswrapper[4838]: I1207 10:36:08.100984 4838 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:36:08 crc kubenswrapper[4838]: I1207 10:36:08.491164 4838 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.110511 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c47cd"] Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.111085 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c47cd" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="registry-server" containerID="cri-o://cc11e4783aba1b2c63333fd803a9796f86fbcea2a2de4b4c634c49665280bae1" gracePeriod=2 Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.469375 4838 generic.go:334] "Generic (PLEG): container finished" podID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerID="cc11e4783aba1b2c63333fd803a9796f86fbcea2a2de4b4c634c49665280bae1" exitCode=0 Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.469465 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47cd" event={"ID":"773fe600-962c-4dc8-acea-b7a25a2a3c38","Type":"ContainerDied","Data":"cc11e4783aba1b2c63333fd803a9796f86fbcea2a2de4b4c634c49665280bae1"} Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.618572 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.685793 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v57h4\" (UniqueName: \"kubernetes.io/projected/773fe600-962c-4dc8-acea-b7a25a2a3c38-kube-api-access-v57h4\") pod \"773fe600-962c-4dc8-acea-b7a25a2a3c38\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.685916 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-catalog-content\") pod \"773fe600-962c-4dc8-acea-b7a25a2a3c38\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.686086 4838 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-utilities\") pod \"773fe600-962c-4dc8-acea-b7a25a2a3c38\" (UID: \"773fe600-962c-4dc8-acea-b7a25a2a3c38\") " Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.687151 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-utilities" (OuterVolumeSpecName: "utilities") pod "773fe600-962c-4dc8-acea-b7a25a2a3c38" (UID: "773fe600-962c-4dc8-acea-b7a25a2a3c38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.694044 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773fe600-962c-4dc8-acea-b7a25a2a3c38-kube-api-access-v57h4" (OuterVolumeSpecName: "kube-api-access-v57h4") pod "773fe600-962c-4dc8-acea-b7a25a2a3c38" (UID: "773fe600-962c-4dc8-acea-b7a25a2a3c38"). InnerVolumeSpecName "kube-api-access-v57h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.741159 4838 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "773fe600-962c-4dc8-acea-b7a25a2a3c38" (UID: "773fe600-962c-4dc8-acea-b7a25a2a3c38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.788139 4838 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v57h4\" (UniqueName: \"kubernetes.io/projected/773fe600-962c-4dc8-acea-b7a25a2a3c38-kube-api-access-v57h4\") on node \"crc\" DevicePath \"\"" Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.788331 4838 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 10:36:11 crc kubenswrapper[4838]: I1207 10:36:11.788413 4838 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773fe600-962c-4dc8-acea-b7a25a2a3c38-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 10:36:12 crc kubenswrapper[4838]: I1207 10:36:12.482325 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c47cd" event={"ID":"773fe600-962c-4dc8-acea-b7a25a2a3c38","Type":"ContainerDied","Data":"480b5a024297248a7812b352e2e48bb556e1933d32731096630e9da5ba84104e"} Dec 07 10:36:12 crc kubenswrapper[4838]: I1207 10:36:12.482660 4838 scope.go:117] "RemoveContainer" containerID="cc11e4783aba1b2c63333fd803a9796f86fbcea2a2de4b4c634c49665280bae1" Dec 07 10:36:12 crc kubenswrapper[4838]: I1207 10:36:12.482881 4838 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c47cd" Dec 07 10:36:12 crc kubenswrapper[4838]: I1207 10:36:12.508061 4838 scope.go:117] "RemoveContainer" containerID="01305b009a24c781dfe4960ebf546ed581bf4467c29250d12d8e2ffa255232be" Dec 07 10:36:12 crc kubenswrapper[4838]: I1207 10:36:12.524967 4838 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c47cd"] Dec 07 10:36:12 crc kubenswrapper[4838]: I1207 10:36:12.531370 4838 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c47cd"] Dec 07 10:36:12 crc kubenswrapper[4838]: I1207 10:36:12.547113 4838 scope.go:117] "RemoveContainer" containerID="2bbcff80c35cb726643d5665c28f94e6da95921329807f8a1404b1ae4470d229" Dec 07 10:36:13 crc kubenswrapper[4838]: I1207 10:36:13.626222 4838 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" path="/var/lib/kubelet/pods/773fe600-962c-4dc8-acea-b7a25a2a3c38/volumes" Dec 07 10:36:54 crc kubenswrapper[4838]: I1207 10:36:54.493624 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:36:54 crc kubenswrapper[4838]: I1207 10:36:54.494800 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:37:24 crc kubenswrapper[4838]: I1207 10:37:24.493387 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:37:24 crc kubenswrapper[4838]: I1207 10:37:24.495279 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:37:54 crc kubenswrapper[4838]: I1207 10:37:54.493681 4838 patch_prober.go:28] interesting pod/machine-config-daemon-7v9qs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 10:37:54 crc kubenswrapper[4838]: I1207 10:37:54.494295 4838 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 10:37:54 crc kubenswrapper[4838]: I1207 10:37:54.494355 4838 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" Dec 07 10:37:54 crc kubenswrapper[4838]: I1207 10:37:54.494938 4838 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2b7f561d1cd9d220ecd241c8416d4d853908ea0e5717620a2f7d502679e5af8"} pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 10:37:54 crc kubenswrapper[4838]: I1207 10:37:54.495007 4838 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" podUID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerName="machine-config-daemon" containerID="cri-o://a2b7f561d1cd9d220ecd241c8416d4d853908ea0e5717620a2f7d502679e5af8" gracePeriod=600 Dec 07 10:37:55 crc kubenswrapper[4838]: I1207 10:37:55.567142 4838 generic.go:334] "Generic (PLEG): container finished" podID="d1304517-3ac2-4aa3-afe2-01e27f87d880" containerID="a2b7f561d1cd9d220ecd241c8416d4d853908ea0e5717620a2f7d502679e5af8" exitCode=0 Dec 07 10:37:55 crc kubenswrapper[4838]: I1207 10:37:55.567216 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerDied","Data":"a2b7f561d1cd9d220ecd241c8416d4d853908ea0e5717620a2f7d502679e5af8"} Dec 07 10:37:55 crc kubenswrapper[4838]: I1207 10:37:55.567769 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7v9qs" event={"ID":"d1304517-3ac2-4aa3-afe2-01e27f87d880","Type":"ContainerStarted","Data":"ff2ec2937f23c705278f544706e9740544c5eef8160464860135423cd05f6603"} Dec 07 10:37:55 crc kubenswrapper[4838]: I1207 10:37:55.567806 4838 scope.go:117] "RemoveContainer" containerID="142390c01a4f70ccedeeb1f2ef820fdaff25330de930d2256a5d42ef8593d81b" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.810312 4838 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cg4f7"] Dec 07 10:39:23 crc kubenswrapper[4838]: E1207 10:39:23.811428 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="extract-utilities" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.811444 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="extract-utilities" Dec 07 10:39:23 crc kubenswrapper[4838]: E1207 10:39:23.811458 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="registry-server" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.811465 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="registry-server" Dec 07 10:39:23 crc kubenswrapper[4838]: E1207 10:39:23.811484 4838 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="extract-content" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.811490 4838 state_mem.go:107] "Deleted CPUSet assignment" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="extract-content" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.811770 4838 memory_manager.go:354] "RemoveStaleState removing state" podUID="773fe600-962c-4dc8-acea-b7a25a2a3c38" containerName="registry-server" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.813188 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.841334 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cg4f7"] Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.895807 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-utilities\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.895916 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpsvs\" (UniqueName: \"kubernetes.io/projected/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-kube-api-access-tpsvs\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.896078 4838 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-catalog-content\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.997335 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpsvs\" (UniqueName: \"kubernetes.io/projected/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-kube-api-access-tpsvs\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.997461 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-catalog-content\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.997531 4838 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-utilities\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.997915 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-catalog-content\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:23 crc kubenswrapper[4838]: I1207 10:39:23.998027 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-utilities\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:24 crc kubenswrapper[4838]: I1207 10:39:24.021928 4838 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpsvs\" (UniqueName: \"kubernetes.io/projected/d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4-kube-api-access-tpsvs\") pod \"community-operators-cg4f7\" (UID: \"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4\") " pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:24 crc kubenswrapper[4838]: I1207 10:39:24.216307 4838 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cg4f7" Dec 07 10:39:24 crc kubenswrapper[4838]: I1207 10:39:24.732175 4838 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cg4f7"] Dec 07 10:39:25 crc kubenswrapper[4838]: I1207 10:39:25.611509 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4f7" event={"ID":"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4","Type":"ContainerDied","Data":"173cfb4f0917ea94baca7fb3730dfc5e9984f09c727cb64b7c2382746871c275"} Dec 07 10:39:25 crc kubenswrapper[4838]: I1207 10:39:25.612384 4838 generic.go:334] "Generic (PLEG): container finished" podID="d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4" containerID="173cfb4f0917ea94baca7fb3730dfc5e9984f09c727cb64b7c2382746871c275" exitCode=0 Dec 07 10:39:25 crc kubenswrapper[4838]: I1207 10:39:25.612427 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4f7" event={"ID":"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4","Type":"ContainerStarted","Data":"11312626260afeb827b367ed81941dda6ce9733b5109c1b38c0057ab6d133364"} Dec 07 10:39:26 crc kubenswrapper[4838]: I1207 10:39:26.622511 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4f7" event={"ID":"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4","Type":"ContainerStarted","Data":"4d78245d5444cd035bfd71a36914baffc1febca32c7186d3c09b6be602ef6830"} Dec 07 10:39:27 crc kubenswrapper[4838]: I1207 10:39:27.648549 4838 generic.go:334] "Generic (PLEG): container finished" podID="d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4" containerID="4d78245d5444cd035bfd71a36914baffc1febca32c7186d3c09b6be602ef6830" exitCode=0 Dec 07 10:39:27 crc kubenswrapper[4838]: I1207 10:39:27.652654 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4f7" event={"ID":"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4","Type":"ContainerDied","Data":"4d78245d5444cd035bfd71a36914baffc1febca32c7186d3c09b6be602ef6830"} Dec 07 10:39:28 crc kubenswrapper[4838]: I1207 10:39:28.661361 4838 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4f7" event={"ID":"d2bad0bb-27e6-4a4c-b173-1db7f0bfa0f4","Type":"ContainerStarted","Data":"9ec3686dc3f8c4ca3fd917004a3680284f8e3d254935f4a6c76e1bde0a726c96"} Dec 07 10:39:28 crc kubenswrapper[4838]: I1207 10:39:28.686937 4838 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cg4f7" podStartSLOduration=3.258512391 podStartE2EDuration="5.686921237s" podCreationTimestamp="2025-12-07 10:39:23 +0000 UTC" firstStartedPulling="2025-12-07 10:39:25.618027219 +0000 UTC m=+5582.325346276" lastFinishedPulling="2025-12-07 10:39:28.046436075 +0000 UTC m=+5584.753755122" observedRunningTime="2025-12-07 10:39:28.678297013 +0000 UTC m=+5585.385616030" watchObservedRunningTime="2025-12-07 10:39:28.686921237 +0000 UTC m=+5585.394240254"